首页 | 官方网站   微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   10403篇
  免费   1472篇
  国内免费   435篇
数理化   12310篇
  2024年   24篇
  2023年   95篇
  2022年   322篇
  2021年   308篇
  2020年   215篇
  2019年   239篇
  2018年   268篇
  2017年   404篇
  2016年   512篇
  2015年   352篇
  2014年   627篇
  2013年   783篇
  2012年   577篇
  2011年   670篇
  2010年   515篇
  2009年   671篇
  2008年   691篇
  2007年   700篇
  2006年   567篇
  2005年   501篇
  2004年   439篇
  2003年   387篇
  2002年   356篇
  2001年   294篇
  2000年   271篇
  1999年   224篇
  1998年   210篇
  1997年   185篇
  1996年   156篇
  1995年   123篇
  1994年   97篇
  1993年   91篇
  1992年   82篇
  1991年   46篇
  1990年   48篇
  1989年   31篇
  1988年   32篇
  1987年   25篇
  1986年   33篇
  1985年   36篇
  1984年   31篇
  1983年   10篇
  1982年   23篇
  1981年   7篇
  1980年   5篇
  1979年   5篇
  1978年   3篇
  1977年   5篇
  1959年   5篇
  1957年   2篇
排序方式: 共有10000条查询结果,搜索用时 0 毫秒
1.
2.
Thin films (monolayer and bilayer) of cylinder forming polystyrene‐block‐polydimethylsiloxane (PS‐b‐PDMS) were shear aligned by the swelling and deswelling of a crosslinked PDMS pad that was physically adhered to the film during solvent vapor annealing. The nanostructures formed by self‐assembly were exposed to ultraviolet‐ozone to partially oxidize the PDMS, followed by calcination in air at 500 °C. In this process, the PS segments were fully decomposed, while the PDMS yielded silica nanostructures. The highly aligned PDMS cylinders were thus deposited as silica nanolines on the silicon substrate. Using a bilayer film, the center‐to‐center distance of these features were effectively halved from 38 to 19 nm. Similarly, by sequential shear‐alignment of two distinct layers, a rhombic array of silica nanolines was fabricated. This methodology provides a facile route to fabricating complex topographically patterned nanostructures. © 2015 Wiley Periodicals, Inc. J. Polym. Sci., Part B: Polym. Phys. 2015 , 53, 1058–1064  相似文献   
3.
Local and Parallel Finite Element Algorithms for Eigenvalue Problems   总被引:4,自引:0,他引:4  
Abstract Some new local and parallel finite element algorithms are proposed and analyzed in this paper foreigenvalue problems.With these algorithms, the solution of an eigenvalue problem on a fine grid is reduced tothe solution of an eigenvalue problem on a relatively coarse grid together with solutions of some linear algebraicsystems on fine grid by using some local and parallel procedure.A theoretical tool for analyzing these algorithmsis some local error estimate that is also obtained in this paper for finite element approximations of eigenvectorson general shape-regular grids.  相似文献   
4.
The affine-scaling modification of Karmarkar's algorithm is extended to solve problems with free variables. This extended primal algorithm is used to prove two important results. First the geometrically elegant feasibility algorithm proposed by Chandru and Kochar is the same algorithm as the one obtained by appending a single column of residuals to the constraint matrix. Second the dual algorithm as first described by Adler et al., is the same as the extended primal algorithm applied to the dual.  相似文献   
5.
In this paper, finite-dimensional recursive filters for space-time Markov random fields are derived. These filters can be used with the expectation maximization (EM) algorithm to yield maximum likelihood estimates of the parameters of the model.  相似文献   
6.
A new method for calculating the radial spheroidal functions of the first kind is proposed for the arguments that are greater than unity in modulus. A well-known representation of these functions is refined and used for this purpose. The constructs and the software implementation proposed in the paper provide an efficient tool for the calculation of the functions with a desired accuracy in a wide range of parameters.  相似文献   
7.
When an organization solves a portfolio problem with public projects evaluated by multiple criteria, in which the economic dimension is not essential or not well characterized, the classical methods are not useful. We propose a non-linear preference model developed from normative Value Theory and using fuzzy sets to model some sources of imprecision. This model can be considered as a generalization of the classical approaches. However, the optimization problem is very complex in order to be solved with non-linear programming techniques. Therefore, the model is exploited by an evolutionary algorithm, able to achieve a strong improvement of the quality of solution.  相似文献   
8.
Toward Fuzzy Optimization without Mathematical Ambiguity   总被引:15,自引:0,他引:15  
Fuzzy programming has been discussed widely in literature and applied in such various disciplines as operations research, economic management, business administration, and engineering. The main purpose of this paper is to present a brief review on fuzzy programming models, and classify them into three broad classes: expected value model, chance-constrained programming and dependent-chance programming. In order to solve general fuzzy programming models, a hybrid intelligent algorithm is also documented. Finally, some related topics are discussed.  相似文献   
9.
讨论了求解无约束线性最小二乘问题的一种并行单纯形法以及对它的改进算法并行共轭梯度—单纯形法 .算法本身具有很强的并行机制 ,能够充分地发挥并行机快速省时的特点 .本文也对算法做了理论分析 ,对算法的收敛性给予了证明 (在二维情形下 ) .最后做了数值实验 (由于软硬件条件的限制 ,并行算法未能在并行计算机上实现 ,鉴于这种情况 ,我们所做的数值实验均是在串行机上完成的 )  相似文献   
10.
On effectiveness of wiretap programs in mapping social networks   总被引:1,自引:0,他引:1  
Snowball sampling methods are known to be a biased toward highly connected actors and consequently produce core-periphery networks when these may not necessarily be present. This leads to a biased perception of the underlying network which can have negative policy consequences, as in the identification of terrorist networks. When snowball sampling is used, the potential overload of the information collection system is a distinct problem due to the exponential growth of the number of suspects to be monitored. In this paper, we focus on evaluating the effectiveness of a wiretapping program in terms of its ability to map the rapidly evolving networks within a covert organization. By running a series of simulation-based experiments, we are able to evaluate a broad spectrum of information gathering regimes based on a consistent set of criteria. We conclude by proposing a set of information gathering programs that achieve higher effectiveness then snowball sampling, and at a lower cost. Maksim Tsvetovat is an Assistant Professor at the Center for Social Complexity and department of Public and International Affairs at George Mason University, Fairfax, VA. He received his Ph.D. from the Computation, Organizations and Society program in the School of Computer Science, Carnegie Mellon University. His dissertation was centered on use of artificial intelligence techniques such as planning and semantic reasoning as a means of studying behavior and evolution of complex social networks, such as these of terrorist organizations. He received a Master of Science degree from University of Minnesota with a specialization in Artificial Intelligence and design of Multi-Agent Systems, and has also extensively studied organization theory and social science research methods. His research is centered on building high-fidelity simulations of social and organizational systems using concepts from distributed artificial intelligence and multi-agent systems. Other projects focus on social network analysis for mapping of internal corporate networks or study of covert and terrorist orgnaizations. Maksim’s vita and publications can be found on Kathleen M. Carley is a professor in the School of Computer Science at Carnegie Mellon University and the director of the center for Compuational Analysis of Social and Organizational Systems (CASOS) which has over 25 members, both students and research staff. Her research combines cognitive science, social networks and computer science to address complex social and organizational problems. Her specific research areas are dynamic network analysis, computational social and organization theory, adaptation and evolution, text mining, and the impact of telecommunication technologies and policy on communication, information diffusion, disease contagion and response within and among groups particularly in disaster or crisis situations. She and her lab have developed infrastructure tools for analyzing large scale dynamic networks and various multi-agent simulation systems. The infrastructure tools include ORA, a statistical toolkit for analyzing and visualizing multi-dimensional networks. ORA results are organized into reports that meet various needs such as the management report, the mental model report, and the intelligence report. Another tool is AutoMap, a text-mining systems for extracting semantic networks from texts and then cross-classifying them using an organizational ontology into the underlying social, knowledge, resource and task networks. Her simulation models meld multi-agent technology with network dynamics and empirical data. Three of the large-scale multi-agent network models she and the CASOS group have developed in the counter-terrorism area are: BioWar a city-scale dynamic-network agent-based model for understanding the spread of disease and illness due to natural epidemics, chemical spills, and weaponized biological attacks; DyNet a model of the change in covert networks, naturally and in response to attacks, under varying levels of information uncertainty; and RTE a model for examining state failure and the escalation of conflict at the city, state, nation, and international as changes occur within and among red, blue, and green forces. She is the founding co-editor with Al. Wallace of the journal Computational Organization Theory and has co-edited several books and written over 100 articles in the computational organizations and dynamic network area. Her publications can be found at: http://www.casos.cs.cmu.edu/bios/carley/publications.php  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号