When.com Web Search

Search results

  1. Results From The WOW.Com Content Network
  2. Residual neural network - Wikipedia

    en.wikipedia.org/wiki/Residual_neural_network

    A residual neural network (also referred to as a residual network or ResNet) [1] is a deep learning architecture in which the layers learn residual functions with reference to the layer inputs. It was developed in 2015 for image recognition , and won the ImageNet Large Scale Visual Recognition Challenge ( ILSVRC ) of that year.

  3. Roofline model - Wikipedia

    en.wikipedia.org/wiki/Roofline_model

    Example of a naïve roofline plot where two kernels are reported. The first (vertical dashed red line) has an arithmetic intensity O 1 {\displaystyle O_{1}} that is underneath the peak bandwidth ceiling (diagonal solid black line), and is then memory-bound .

  4. Torch (machine learning) - Wikipedia

    en.wikipedia.org/wiki/Torch_(machine_learning)

    The torch.class(classname, parentclass) function can be used to create object factories . When the constructor is called, torch initializes and sets a Lua table with the user-defined metatable , which makes the table an object .

  5. Test functions for optimization - Wikipedia

    en.wikipedia.org/wiki/Test_functions_for...

    Here some test functions are presented with the aim of giving an idea about the different situations that optimization algorithms have to face when coping with these kinds of problems. In the first part, some objective functions for single-objective optimization cases are presented.

  6. Bode plot - Wikipedia

    en.wikipedia.org/wiki/Bode_plot

    Figures 2-5 further illustrate construction of Bode plots. This example with both a pole and a zero shows how to use superposition. To begin, the components are presented separately. Figure 2 shows the Bode magnitude plot for a zero and a low-pass pole, and compares the two with the Bode straight line plots.

  7. Line integral convolution - Wikipedia

    en.wikipedia.org/wiki/Line_integral_convolution

    Oriented Line Integral Convolution (OLIC) solves this issue by using a ramp-like asymmetric kernel and a low-density noise texture. [8] The kernel asymmetrically modulates the intensity along the streamline, producing a trace that encodes orientation; the low-density of the noise texture prevents smeared traces from overlapping, aiding readability.