-
机器人 编队 蜂拥 matlab 仿真
该程序用于Matlab机器人编队仿真,主要包括传感器模块,通信模块,控制器模块,感知环境模块,易于修改并使用!
- 2020-12-05下载
- 积分:1
-
二级倒立摆,matlab仿真,simulink建模仿真,lqr最优控制
绝对可以用的二级倒立摆模型。simulink建模,matlab编写s函数,使用lqr最优控制
- 2020-12-07下载
- 积分:1
-
MIPI M-PHY.pdf
在MIPI目前公布的协议中,有3类基于摄像头的接口,一个是前几年大行其道的D-PHY接口,一个是C-PHY接口,还有一个是M-PHY接口,这个文档讲的就是M-PHY
- 2021-05-06下载
- 积分:1
-
动态规划法、贪心算法、回溯法、分支限界法解决0-1背包
1) 动态规划法求解问题的一般思路,动态规划法求解本问题的思路及其C/C++程序实现与算法的效率分析。2) 贪心算法在0-1背包问题求解中的应用3) 回溯法求解问题的一般思路,回溯法求解本问题的思路及其C/C++程序实现与算法的效率分析。4) 分支限界法求解问题的一般思路,分支限界法求解本问题的思路及其C/C++程序实现与算法的效率分析。有代码!!
- 2020-12-02下载
- 积分:1
-
乒乓球比赛抽签软件,含说明书、学习包
乒乓球比赛抽签软件,含说明书、学习包,很好用,手机端也可以使用
- 2020-12-10下载
- 积分:1
-
UKF的MATLAB程序
该程序有详细的注释,易于读者理解与运用。
- 2020-12-07下载
- 积分:1
-
FPGA 做的100M 频率计
基于FPGA的100M频率计设计功能描述: 该频率计是以FPGA为核心器件,嵌入mc8051 IP核,并以整形电路、1602液晶显示器等作为外围设计而成的等精度频率计。通过1602液晶显示被测频率值、周期、脉宽、占空比,闸门时间在0.1—10S连续可调,测量范围为0.1Hz—100MHz。
- 2020-12-03下载
- 积分:1
-
【PDF】《Machine learning A Probabilistic Perspective》 MLAPP;by Kevin Murphy
完整版,带目录,机器学习必备经典;大部头要用力啃。Machine learning A Probabilistic PerspectiveMachine LearningA Probabilistic PerspectiveKevin P. MurphyThe mit PressCambridge, MassachusettsLondon, Englando 2012 Massachusetts Institute of TechnologyAll rights reserved. No part of this book may be reproduced in any form by any electronic or mechanicalmeans(including photocopying, recording, or information storage and retrieval)without permission inwriting from the publisherFor information about special quantity discounts, please email special_sales@mitpress. mit. eduThis book was set in the HEx programming language by the author. Printed and bound in the UnitedStates of AmLibrary of Congress Cataloging-in-Publication InformationMurphy, Kevin Png:a piobabilistctive/Kevin P. Murphyp. cm. -(Adaptive computation and machine learning series)Includes bibliographical references and indexisBn 978-0-262-01802-9 (hardcover: alk. paper1. Machine learning. 2. Probabilities. I. TitleQ325.5M872012006.31-dc232012004558109876This book is dedicated to alessandro, Michael and stefanoand to the memory of gerard Joseph murphyContentsPreactXXVII1 IntroductionMachine learning: what and why?1..1Types of machine learning1.2 Supervised learning1.2.1Classification 31.2.2 Regression 83 Unsupervised learning 91.3.11.3.2Discovering latent factors 111.3.3 Discovering graph structure 131.3.4 Matrix completion 141.4 Some basic concepts in machine learning 161.4.1Parametric vs non-parametric models 161.4.2 A simple non-parametric classifier: K-nearest neighbors 161.4.3 The curse of dimensionality 181.4.4 Parametric models for classification and regression 191.4.5Linear regression 191.4.6Logistic regression1.4.7 Overfitting 221.4.8Model selection1.4.9No free lunch theorem242 Probability2.1 Introduction 272.2 A brief review of probability theory 282. 2. 1 Discrete random variables 282. 2.2 Fundamental rules 282.2.3B292. 2. 4 Independence and conditional independence 302. 2. 5 Continuous random variable32CONTENTS2.2.6 Quantiles 332.2.7 Mean and variance 332.3 Some common discrete distributions 342.3.1The binomial and bernoulli distributions 342.3.2 The multinomial and multinoulli distributions 352. 3.3 The Poisson distribution 372.3.4 The empirical distribution 372.4 Some common continuous distributions 382.4.1 Gaussian (normal) distribution 382.4.2Dte pdf 392.4.3 The Laplace distribution 412.4.4 The gamma distribution 412.4.5 The beta distribution 422.4.6 Pareto distribution2.5 Joint probability distributions 442.5.1Covariance and correlation442.5.2 The multivariate gaussian2.5.3 Multivariate Student t distribution 462.5.4 Dirichlet distribution 472.6 Transformations of random variables 492. 6. 1 Linear transformations 492.6.2 General transformations 502.6.3 Central limit theorem 512.7 Monte Carlo approximation 522.7.1 Example: change of variables, the MC way 532.7.2 Example: estimating T by Monte Carlo integration2.7.3 Accuracy of Monte Carlo approximation 542.8 Information theory562.8.1Entropy2.8.2 KL dive572.8.3 Mutual information 593 Generative models for discrete data 653.1 Introducti653.2 Bayesian concept learning 653.2.1Likelihood673.2.2 Prior 673.2.3P683.2.4Postedictive distribution3.2.5 A more complex prior 723.3 The beta-binomial model 723.3.1 Likelihood 733.3.2Prior743.3.3 Poster3.3.4Posterior predictive distributionCONTENTS3.4 The Dirichlet-multinomial model 783. 4. 1 Likelihood 793.4.2 Prior 793.4.3 Posterior 793.4.4Posterior predictive813.5 Naive Bayes classifiers 823.5.1 Model fitting 833.5.2 Using the model for prediction 853.5.3 The log-sum-exp trick 803.5.4 Feature selection using mutual information 863.5.5 Classifying documents using bag of words 84 Gaussian models4.1 Introduction974.1.1Notation974. 1.2 Basics 974. 1.3 MlE for an mvn 994.1.4 Maximum entropy derivation of the gaussian 1014.2 Gaussian discriminant analysis 1014.2.1 Quadratic discriminant analysis(QDA) 1024.2.2 Linear discriminant analysis (LDA) 1034.2.3 Two-claSs LDA 1044.2.4 MLE for discriminant analysis 1064.2.5 Strategies for preventing overfitting 1064.2.6 Regularized LDA* 104.2.7 Diagonal LDA4.2.8 Nearest shrunken centroids classifier1094.3 Inference in jointly Gaussian distributions 1104.3.1Statement of the result 1114.3.2 Examples4.3.3 Information form 1154.3.4 Proof of the result 1164.4 Linear Gaussian systems 1194.4.1Statement of the result 1194.4.2 Examples 1204.4.3 Proof of the result1244.5 Digression: The Wishart distribution4.5. 1 Inverse Wishart distribution 1264.5.2 Visualizing the wishart distribution* 1274.6 Inferring the parameters of an MVn 1274.6.1 Posterior distribution of u 1284.6.2 Posterior distribution of e1284.6.3 Posterior distribution of u and 2* 1324.6.4 Sensor fusion with unknown precisions 138
- 2020-12-10下载
- 积分:1
-
html2canvas加上canvas2image保存网页为图片
通过纯JS插件对浏览器端的页面进行截图,截图之后再进行保存下载。
- 2020-11-27下载
- 积分:1
-
非线性方程组求解matlab程序
mulStablePoint 用不动点迭代法求非线性方程组的一个根mulNewton 用牛顿法法求非线性方程组的一个根mulDiscNewton 用离散牛顿法法求非线性方程组的一个根mulMix 用牛顿-雅可比迭代法求非线性方程组的一个根mulNewtonSOR 用牛顿-SOR迭代法求非线性方程组的一个根mulDNewton 用牛顿下山法求非线性方程组的一个根mulGXF1 用两点割线法的第一种形式求非线性方程组的一个根mulGXF2 用两点割线法的第二种形式求非线性方程组的一个根mulVNewton 用拟牛顿法求非线性方程组的一
- 2020-12-06下载
- 积分:1