首页
学习
活动
专区
工具
TVP
发布

翻译scikit-learn Cookbook

学习sklearn
专栏作者
78
文章
52734
阅读量
15
订阅数
机器学习重新思考--何为机器学习
在科技发展的今天,很多事情人们都不再亲力亲为,要么交给了专业人士,要么交给了机器。于是有一部分人称为了专业人士,一部分人成了专业忽悠,一部分人去捣鼓机器,然后一大部分人被解放出来,获得了休闲。
到不了的都叫做远方
2020-05-08
3350
数据结构学习-python实现01--0401
经过近两年多的转行自学,乱七八糟的学了不少的东西,依然没有走到自己想要去的方向,继续学习,努力吧!
到不了的都叫做远方
2020-04-01
4400
Classifying documents with Naïve Bayes使用朴素贝叶斯分类文本
Naïve Bayes is a really interesting model. It's somewhat similar to k-NN in the sense that it makes some assumptions that might oversimplify reality, but still perform well in many cases.
到不了的都叫做远方
2019-12-07
3860
Using many Decision Trees – random forests使用多棵决策树--随机森林
In this recipe, we'll use random forests for classification tasks. random forests are used because they're very robust to overfitting and perform well in a variety of situations.
到不了的都叫做远方
2019-11-29
6200
Tuning a Decision Tree model调试决策树模型
If we use just the basic implementation of a Decision Tree, it will probably not fit very well.Therefore, we need to tweak the parameters in order to get a good fit. This is very easy and won't require much effort.
到不了的都叫做远方
2019-11-28
1.2K0
Doing basic classifications with Decision Trees使用决策树做基本分类
In this recipe, we will perform basic classifications using Decision Trees. These are very nice models because they are easily understandable, and once trained in, scoring is very simple.
到不了的都叫做远方
2019-11-27
3560
4 Classifying Data with scikit-learn使用scikit-learn分类数据
This chapter will cover the following topics:本章将涵盖以下主题:
到不了的都叫做远方
2019-11-27
3190
Finding the closest objects in the feature space在特征空间中找到最接近的对象
Sometimes, the easiest thing to do is to just find the distance between two objects. We just need to find some distance metric, compute the pairwise distances, and compare the outcomes to what's expected.
到不了的都叫做远方
2019-11-24
6480
Directly applying Bayesian ridge regression直接使用贝叶斯岭回归
In the Using ridge regression to overcome linear regression's shortfalls recipe, we discussed the connections between the constraints imposed by ridge regression from an optimization standpoint. We also discussed the Bayesian interpretation of priors on the coefficients, which attract the mass of the density towards the prior, which often has a mean of 0 .
到不了的都叫做远方
2019-11-18
1.5K0
Optimizing the ridge regression parameter最优化岭回归参数
Once you start using ridge regression to make predictions or learn about relationships in the system you're modeling, you'll start thinking about the choice of alpha.For example, using OLS regression might show some relationship between two variables;however, when regularized by some alpha, the relationship is no longer significant. This can be a matter of whether a decision needs to be taken.
到不了的都叫做远方
2019-11-13
1.5K0
2 Working with Linear Models 2 线性模型
In this chapter, we will cover the following topics:在这章,将涵盖以下主题:
到不了的都叫做远方
2019-11-10
4380
Using stochastic gradient descent for regression使用随机梯度下降进行回归分析
In this recipe, we'll get our first taste of stochastic gradient descent. We'll use it for regression here, but for the next recipe, we'll use it for classification.
到不了的都叫做远方
2019-11-09
5400
Using Gaussian processes for regression降维之高斯过程
In this recipe, we'll use the Gaussian process for regression. In the linear models section,we saw how representing prior information on the coefficients was possible using Bayesian Ridge Regression.
到不了的都叫做远方
2019-11-07
9750
Imputing missing values through various strategies填充处理缺失值的不同方法
Data imputation is critical in practice, and thankfully there are many ways to deal with it.In this recipe, we'll look at a few of the strategies. However, be aware that there might be other approaches that fit your situation better.
到不了的都叫做远方
2019-10-30
8340
Working with categorical variables处理分类变量
Categorical variables are a problem. On one hand they provide valuable information; on the other hand, it's probably text—either the actual text or integers corresponding to the text—like an index in a lookup table.So, we clearly need to represent our text as integers for the model's sake, but we can't just use the id field or naively represent them. This is because we need to avoid a similar problem to the Creating binary features through thresholding recipe. If we treat data that is continuous, it must be interpreted as continuous.
到不了的都叫做远方
2019-10-29
8050
Creating binary features through thresholding通过阈值来生成二元特征
In the last recipe, we looked at transforming our data into the standard normal distribution.Now, we'll talk about another transformation, one that is quite different.
到不了的都叫做远方
2019-10-28
4190
scikit-learn Cookbook 00
This chapter discusses setting data, preparing data, and premodel dimensionality reduction.These are not the
到不了的都叫做远方
2019-10-25
4120
没有更多了
社区活动
腾讯技术创作狂欢月
“码”上创作 21 天,分 10000 元奖品池!
Python精品学习库
代码在线跑,知识轻松学
博客搬家 | 分享价值百万资源包
自行/邀约他人一键搬运博客,速成社区影响力并领取好礼
技术创作特训营·精选知识专栏
往期视频·千货材料·成员作品 最新动态
领券
问题归档专栏文章快讯文章归档关键词归档开发者手册归档开发者手册 Section 归档