首页 | 官方网站   微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   7368篇
  免费   501篇
  国内免费   10篇
工业技术   7879篇
  2024年   8篇
  2023年   77篇
  2022年   319篇
  2021年   332篇
  2020年   215篇
  2019年   226篇
  2018年   260篇
  2017年   289篇
  2016年   329篇
  2015年   227篇
  2014年   381篇
  2013年   605篇
  2012年   494篇
  2011年   595篇
  2010年   464篇
  2009年   465篇
  2008年   436篇
  2007年   366篇
  2006年   284篇
  2005年   217篇
  2004年   190篇
  2003年   165篇
  2002年   182篇
  2001年   97篇
  2000年   74篇
  1999年   77篇
  1998年   72篇
  1997年   44篇
  1996年   40篇
  1995年   45篇
  1994年   31篇
  1993年   29篇
  1992年   25篇
  1991年   13篇
  1990年   16篇
  1989年   19篇
  1988年   13篇
  1987年   10篇
  1986年   14篇
  1985年   18篇
  1984年   18篇
  1983年   8篇
  1982年   18篇
  1981年   10篇
  1980年   11篇
  1979年   9篇
  1978年   10篇
  1977年   10篇
  1975年   6篇
  1973年   5篇
排序方式: 共有7879条查询结果,搜索用时 15 毫秒
91.
92.
A Wireless Mesh Network (WMN) is composed of multiple Access Points (APs) that are connected together using the radio channel and by a limited number of gateway APs connected to the Internet. In this paper, we address the problem of gateway placement that consists of minimizing the number of gateways while satisfying system performance requirements. Along with the placement problem, the formulation includes joint routing and scheduling to account for the problem of interference and to enable spacial reuse. The problem, which we coined GPSRP (Gateway Placement and Spatial Reuse Problem), allows a much more efficient use of the available resources and reduces overall gateway costs. This article presents for the first time a mathematical formulation of the problem and discusses its advantages and limitations with respect to other approaches.  相似文献   
93.
The availability of multicore processors and programmable NICs, such as TOEs (TCP/IP Offloading Engines), provides new opportunities for designing efficient network interfaces to cope with the gap between the improvement rates of link bandwidths and microprocessor performance. This gap poses important challenges related with the high computational requirements associated to the traffic volumes and wider functionality that the network interface has to support. This way, taking into account the rate of link bandwidth improvement and the ever changing and increasing application demands, efficient network interface architectures require scalability and flexibility. An opportunity to reach these goals comes from the exploitation of the parallelism in the communication path by distributing the protocol processing work across processors which are available in the computer, i.e. multicore microprocessors and programmable NICs.Thus, after a brief review of the different solutions that have been previously proposed for speeding up network interfaces, this paper analyzes the onloading and offloading alternatives. Both strategies try to release host CPU cycles by taking advantage of the communication workload execution in other processors present in the node. Nevertheless, whereas onloading uses another general-purpose processor, either included in a chip multiprocessor (CMP) or in a symmetric multiprocessor (SMP), offloading takes advantage of processors in programmable network interface cards (NICs). From our experiments, implemented by using a full-system simulator, we provide a fair and more complete comparison between onloading and offloading. Thus, it is shown that the relative improvement on peak throughput offered by offloading and onloading depends on the rate of application workload to communication overhead, the message sizes, and on the characteristics of the system architecture, more specifically the bandwidth of the buses and the way the NIC is connected to the system processor and memory. In our implementations, offloading provides lower latencies than onloading, although the CPU utilization and interrupts are lower for onloading. Taking into account the conclusions of our experimental results, we propose a hybrid network interface that can take advantage of both, programmable NICs and multicore processors.  相似文献   
94.
In last years, Face recognition based on 3D techniques is an emergent technology which has demonstrated better results than conventional 2D approaches. Using texture (180° multi-view image) and depth maps is supposed to increase the robustness towards the two main challenges in Face Recognition: Pose and illumination. Nevertheless, 3D data should be acquired under highly controlled conditions and in most cases depends on the collaboration of the subject to be recognized. Thus, in applications such as surveillance or control access points, this kind of 3D data may not be available during the recognition process. This leads to a new paradigm using some mixed 2D-3D face recognition systems where 3D data is used in the training but either 2D or 3D information can be used in the recognition depending on the scenario. Following this concept, where only part of the information (partial concept) is used in the recognition, a novel method is presented in this work. This has been called Partial Principal Component Analysis (P2CA) since they fuse the Partial concept with the fundamentals of the well known PCA algorithm. This strategy has been proven to be very robust in pose variation scenarios showing that the 3D training process retains all the spatial information of the face while the 2D picture effectively recovers the face information from the available data. Furthermore, in this work, a novel approach for the automatic creation of 180° aligned cylindrical projected face images using nine different views is presented. These face images are created by using a cylindrical approximation for the real object surface. The alignment is done by applying first a global 2D affine transformation of the image, and afterward a local transformation of the desired face features using a triangle mesh. This local alignment allows a closer look to the feature properties and not the differences. Finally, these aligned face images are used for training a pose invariant face recognition approach (P2CA).  相似文献   
95.
High performance processor designs have evolved toward architectures that integrate multiple processing cores on the same chip. As the number of cores inside a Chip MultiProcessor (CMP) increases, the interconnection network will have significant impact on both overall performance and energy consumption as previous studies have shown. Moreover, wires used in such interconnect can be designed with varying latency, bandwidth and power characteristics. In this work, we show how messages can be efficiently managed in tiled CMP, from the point of view of both performance and energy, by combining both address compression with a heterogeneous interconnect. In particular, our proposal is based on applying an address compression scheme that dynamically compresses the addresses within coherence messages allowing for a significant area slack. The arising area is exploited for wire latency improvement by using a heterogeneous interconnection network comprised of a small set of very-low-latency wires for critical short-messages in addition to baseline wires. Detailed simulations of a 16-core CMP show that our proposal obtains average improvements of 10% in execution time and 38% in the energy-delay2 product of the interconnect. Additionally, the sensitivity analysis shows that our proposal performs well when either OoO cores or caches with higher latencies are considered.  相似文献   
96.
We demonstrate a technique to recirculate liquids in a microfluidic channel by alternating predominance of centrifugal and capillary forces to rapidly bring the entire volume of a liquid sample to within one diffusion length, δ, of the surface, even for sample volumes hundreds of times the product of δ and the geometric device area. This is accomplished by repetitive, random sampling of an on-disc sample reservoir to form a thin fluid layer of thickness δ in a microchannel, maintaining contact for the diffusion time, then rapidly exchanging the fluid layer for a fresh aliquot by disc rotation and stoppage. With this technique, liquid volumes of microlitres to millilitres can be handled in many sizes of microfluidic channels, provided the channel wall with greatest surface area is hydrophilic. We present a theoretical model describing the balance of centrifugal and capillary forces in the device and validate the model experimentally.  相似文献   
97.
The advances in the educational field and the high complexity of student modeling have provoked it to be one of the aspects more investigated in Intelligent Tutoring Systems (ITSs). The Student Models (SMs) should not only represent the student’s knowledge, but rather they should reflect, as faithfully as possible, the student’s reasoning process. To facilitate this goal, in this article a new approach to student modeling is proposed that benefits from the advantages of Ontological Engineering, advancing in the pursue of a more granular and complete knowledge representation. It’s focused, mainly, on the SM cognitive diagnosis process, and we present a method providing a rich diagnosis about the student’s knowledge state – especially, about the state of learning objectives reached or not. The main goal is to achieve SMs with a good adaptability to the student’s features and a high flexibility for its integration in varied ITSs.  相似文献   
98.
The IEEE 802.21 standard facilitates media independent handovers by providing higher layer mobility management functions with common service primitives for all technologies. Right after the base specification was published, several voices rose up in the working group advocating to broaden the scope of IEEE 802.21 beyond handovers. This paper aims at updating the reader with the main challenges and functionalities required to create a Media Independence Service Layer, through the analysis of scenarios which are being discussed within the working group: 1) Wireless Coexistence, and 2) Heterogeneous Wireless Multihop Backhaul Networks.  相似文献   
99.
Massive convolution is the basic operation in multichannel acoustic signal processing. This field has experienced a major development in recent years. One reason for this has been the increase in the number of sound sources used in playback applications available to users. Another reason is the growing need to incorporate new effects and to improve the hearing experience. Massive convolution requires high computing capacity. GPUs offer the possibility of parallelizing these operations. This allows us to obtain the processing result in much shorter time and to free up CPU resources. One important aspect lies in the possibility of overlapping the transfer of data from CPU to GPU and vice versa with the computation, in order to carry out real-time applications. Thus, a synthesis of 3D sound scenes could be achieved with only a peer-to-peer music streaming environment using a simple GPU in your computer, while the CPU in the computer is being used for other tasks. Nowadays, these effects are obtained in theaters or funfairs at a very high cost, requiring a large quantity of resources. Thus, our work focuses on two mains points: to describe an efficient massive convolution implementation and to incorporate this task to real-time multichannel-sound applications.  相似文献   
100.
In this paper, we propose a methodology for training a new model of artificial neural network called the generalized radial basis function (GRBF) neural network. This model is based on generalized Gaussian distribution, which parametrizes the Gaussian distribution by adding a new parameter τ. The generalized radial basis function allows different radial basis functions to be represented by updating the new parameter τ. For example, when GRBF takes a value of τ=2, it represents the standard Gaussian radial basis function. The model parameters are optimized through a modified version of the extreme learning machine (ELM) algorithm. In the methodology proposed (MELM-GRBF), the centers of each GRBF were taken randomly from the patterns of the training set and the radius and τ values were determined analytically, taking into account that the model must fulfil two constraints: locality and coverage. An thorough experimental study is presented to test its overall performance. Fifteen datasets were considered, including binary and multi-class problems, all of them taken from the UCI repository. The MELM-GRBF was compared to ELM with sigmoidal, hard-limit, triangular basis and radial basis functions in the hidden layer and to the ELM-RBF methodology proposed by Huang et al. (2004) [1]. The MELM-GRBF obtained better results in accuracy than the corresponding sigmoidal, hard-limit, triangular basis and radial basis functions for almost all datasets, producing the highest mean accuracy rank when compared with these other basis functions for all datasets.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号