搜索资源列表
descent
- 自适应滤波中的最速下降法仿真程序,课程的大作业,可以绘制梯度曲线-Adaptive Filter steepest descent method simulation program, course the big job, you can draw the gradient curve
hmmtrain
- 1.自己推导出来的mean与covariance来train theta 2.自己推导出来的mean与covariance来跑Viterbi 3.推theta时的Gradient Descent: Learing Rate: 0.000001 GD iter : 50 iteration 4.Scaler : 1/260-1. Their derived mean and covariance to train theta 2. Their derive
DE
- 差分进化算法(Differential Evolution, DE)是一种新兴的进化计算技术。它是由Storn等人于1995年提出的,最初的设想是用于解决切比雪夫多项式问题,后来发现DE也是解决复杂优化问题的有效技术。本人实现了简单的DE-DE is used for multidimensional real-valued functions but does not use the gradient of the problem being optimized, which means DE
BP1
- 采用动量梯度下降算法训练 BP 神经网络预测的一个实例分析-Gradient descent algorithm with momentum training BP neural network analysis of an instance of
matlab
- 最速下降梯度法matlab程序和用拟牛顿法求极小值点-apply a gradient–descent based error minimization approach to minimize the difference between the real deformation and the simulated one given the elasticity parameters
BP-NET
- 用动量梯度下降法训练BP网络 已知输入向量为P=[-1,-2,3,1 -1,1,5,3],目标输出为T=[-1,-1,1,1]。 -Gradient descent with momentum BP network training input vector is known as P = [-1,-2,3,1 -1,1,5,3], the target output for the T = [-1,-1,1, 1].
DescenteGradient
- gradient descent very good document.
BP-Iris-classifier
- 使用BP网络实现了对Iris数据的分类,使用了可变学习速率和带动量的梯度下降算法。-Using the BP network realizes the classification of Iris data, the use of the variable learning rate and the amount of gradient descent algorithm driven.
Widrow-Hoff-Training-Method
- The WH class is an implementation of the Widrow-Hoff. The Widrow-Hoff (WH) algorithm, often called Least Mean Square (LMS), is an online-algorithm. The WH can be interpreted as a gradient descent procedure on the error space. In other words, the
Conjugate-gradient
- 共轭梯度法是介于最速下降法与牛顿法之间的一个方法,它仅需利用一阶导数信息,但克服了最速下降法收敛慢的缺点,又避免了牛顿法需要存储和计算Hesse矩阵并求逆的缺点,共轭梯度法不仅是解决大型线性方程组最有用的方法之一,也是解大型非线性最优化最有效的算法之一。-Conjugate gradient method is between the steepest descent method and Newton method between a method that only use the firs
pr2
- 模式识别作业 梯度下降法的java实现代码-Pattern Recognition Task gradient descent implementation code java
simple-grad-desend
- 简单梯度下降法,用于迭代求极值,是一种最速下降法-Simple gradient descent method for the iterative extremum is a steepest descent method
duotezhengzishiyingyuzhi
- 该程序实现多特征自适应阈值算法,两个matlab文件,一个为函数,该函数求出信号的特征量值并通过梯度下降法训练出其初始阈值,另一个为主函数,调用该函数文件;该程序将4路信号,每一路信号的特征量值及其自适应阈值求出来,并一起画出图来。-The program features multi-adaptive threshold algorithm, two matlab files, one for the function,the function seek out characteristics
gaijinzishiyingxiangguan
- 该程序实现改进的自适应相关算法,两个matlab文件,一个为函数,该函数求出信号特征量的马氏距离平方并通过梯度下降法训练出其固定阈值,另一个为主函数,调用该函数文件;该程序将4路信号,每一路信号特征量的马氏距离平方及其固定阈值求出来,并一起画出图来。 -The program features improved adaptive correlation algorithm, two matlab files, one for the function that calculate the
bp1
- bp神经网络算法大全,包括基本算法、最优梯度下降算法、自适应算法等-bp neural network algorithm Daquan, including the basic algorithm, the optimal gradient descent algorithm, adaptive algorithm
BPANN
- 本例采用vc编程实现BP神经网络,程序将涉及到梯度下降法、函数求逆,以及偏导数的实现。-In this case using vc BP neural network programming, the program will involve the gradient descent method, the inverse function, as well as the realization of partial derivatives.
gradientDescent.m
- gradient descent in matlab
xitongbianshi
- 系统辨识基于梯度校正的参数辨识方法,其中包括随机牛顿法,确定性梯度下降法-System identification based on the gradient correction parameter identification methods, including random Newton' s method, deterministic gradient descent
machinelearning
- 神经网络 解决1d marble ,用 gradient descent 来训练神经网络-neural network solving 1d marble problem
Establishment-of-BP-neural-network
- 构建BP神经网络,使用梯度下降法训练BP神经网络-Establishment of BP neural network using gradient descent training of BP neural network