当前位置:网站首页>Deep learning, thinking from one dimensional input to multi-dimensional feature input
Deep learning, thinking from one dimensional input to multi-dimensional feature input
2022-07-03 05:48:00 【code bean】
From one dimensional input to multidimensional features
When looking at one dimension alone , I haven't felt anything yet .
But when I see this picture , My first reaction was , The original meaning of activation function is !


( But here is another point to note : Here is only the input dimension increased , Output or 0 and 1, So this is still a problem of two categories , Only the input characteristics become multidimensional , When it comes to multi classification, we should introduce softmax 了 )
At first I was thinking , Why should we add the eight characteristics of a piece of data into a linear equation Give again “ Sigma ” function ?
Then think of , Because each feature contributes differently to the whole , So there are different wb, Yes wb Blessing , So adding eigenvalues is not incomprehensible . then Linear systems The result value of “ Sigma ” function , In this way, the system has nonlinearity , This is also the purpose of the activation function —— It can fit more complex decision boundaries .
Then look at this picture :

Two implications of this picture
1 All forms of multiplication and accumulation , In fact, it can be converted into Multiplication of vectors , In turn to , Why should vector multiplication be designed like this , It may be to simplify the writing of accumulation .
2 After being written as a vector, you can use many vector properties . such as , It should have been x*w, Now I want to become w*x, But matrices have no commutative law , But according to the characteristics of transpose , We can xw All transposed , So they can change positions .
But there are still problems , there b How did you get it ? From the formula, it seems to appear out of thin air !
So I feel that there is a little problem with this way of writing , The more common way to write it should be , Add a dimension :

That is, first in X Add a... In front of the 1, As the base of the offset term ,( here X From n The dimensional vector becomes n+1 Dimension vector , Turn into [1, x1,x2…] ), then , Let each classifier train its own bias term weight , So the weight of each classifier becomes n+1 dimension , namely [w0,w1,…], among ,w0 Is the weight of the offset term , therefore 1*w0 Is the bias of this classifier / Intercept . such , Just let the intercept b This seems to be related to the slope W Different parameters , It's all unified to the next framework , Make the model constantly adjust parameters in the process of training w0, So as to achieve adjustment b Purpose .
————————————————
Copyright notice : This paper is about CSDN Blogger 「 Almost Human V」 The original article of , follow CC 4.0 BY-SA Copyright agreement , For reprint, please attach the original source link and this statement .
Link to the original text :https://blog.csdn.net/Uwr44UOuQcNsUQb60zk2/article/details/81074408
But I still have a problem , If in ,X Add a... In front of the 1 For the base , But adjust w0 When , The base is 1, That's equivalent to every linear function b It's still the same ? Then if it is here 1 Change to the previous linear function b Is it more reasonable ?
I'm not sure here ( Follow up in understanding .)
In a word, this offset term b The role of is Build in space No Across the origin A straight line / Plane / hyperplane . In this way, we can better classify ( Better build decision boundaries )( Digression : But in Euclidean space vectors, the vectors we discuss all cross the origin )


Solved the case
The lines here are not the origin , But vectors in Euclidean space all cross the origin , How do these vectors come from , I don't know if you have the same question as me .
This is before me , A question raised :
After communicating with the teacher , My conclusion is that :
That can be understood from two aspects :
The first one is :
A one-dimensional Euclidean space X Store all x The possibility of , A one-dimensional Euclidean space Y Store all y The possibility of .
And in two Euclidean spaces xy There is a one-to-one correspondence , With this relationship ,
Each pair xy It can also form a new two-dimensional Euclidean space XY. If there is another one-dimensional Euclidean space B,
Store all intercept b, And this B and XY There is also a one-to-one correspondence , According to these relationships , We can describe the straight line of intercept .
such as XY There's an element in (x1, y1) B There's an element in b1, Through these three pieces of information, we can describe a line segment with intercept .
And this B It can also be used as a feature item , Put in X Matrix , A new column .The second kind :
Regard Euclidean space as a set of points , stay X and W Before expansion ,XY The space formed by all the points in is actually a straight line passing through the origin .
stay X and W After expanding one dimension ,XY The space formed by all points in is an intercept ( But the origin ) The straight line of .
( you 're right , I was really going to X Y W These spaces are confused , The teacher woke me up here !!!)summary :
Although the vectors in Euclidean space are all vectors passing through the origin , But if you take these vectors as a point ( Ignore origin ), These points constitute
A space . This space might be N A line in dimensional space , One face , Or a person This is the concept of subspace .
That's it , I hope I can help people who are as confused as me .
Reference material :
《PyTorch Deep learning practice 》 Complete the collection _ Bili, Bili _bilibili
Bias term in neural network b What is it ?_ Almost Human V The blog of -CSDN Blog
边栏推荐
- 求质数的方法
- How does win7 solve the problem that telnet is not an internal or external command
- 理解 YOLOV1 第一篇 预测阶段
- Redis cannot connect remotely.
- [branch and cycle] | | super long detailed explanation + code analysis + a trick game
- 伯努利分布,二项分布和泊松分布以及最大似然之间的关系(未完成)
- "C and pointer" - Chapter 13 advanced pointer int * (* (* (*f) () [6]) ()
- [video of Teacher Zhao Yuqiang's speech on wot] redis high performance cache and persistence
- QT read write excel -- qxlsx insert chart 5
- Configure DTD of XML file
猜你喜欢

Redis使用Lua脚本简介

Life is a process of continuous learning

今天很多 CTO 都是被幹掉的,因為他沒有成就業務

Why is the website slow to open?
![[branch and cycle] | | super long detailed explanation + code analysis + a trick game](/img/aa/543d4f0dcbcd664be963579af77ec9.jpg)
[branch and cycle] | | super long detailed explanation + code analysis + a trick game

期末复习(Day5)

一起上水碩系列】Day 9

pytorch DataLoader实现miniBatch(未完成)
![[teacher Zhao Yuqiang] index in mongodb (Part 1)](/img/2d/277ec737f2a7065831a19d036e61e1.jpg)
[teacher Zhao Yuqiang] index in mongodb (Part 1)

Altaro o365 total backup subscription plan
随机推荐
[teacher Zhao Yuqiang] Cassandra foundation of NoSQL database
Get and monitor remote server logs
How to use source insight
牛客网 JS 分隔符
[video of Teacher Zhao Yuqiang's speech on wot] redis high performance cache and persistence
[advanced pointer (1)] | detailed explanation of character pointer, pointer array, array pointer
Solve the 1251 client does not support authentication protocol error of Navicat for MySQL connection MySQL 8.0.11
How to create and configure ZABBIX
Btrfs and ext4 - features, strengths and weaknesses
2022.DAY592
C 语言文件操作函数大全 (超详细)
kubernetes资源对象介绍及常用命令(五)-(ConfigMap)
QT read write excel -- qxlsx insert chart 5
Ansible firewall firewalld setting
Final review (Day6)
The request database reported an error: "could not extract resultset; SQL [n/a]; needed exception is org.hibernate.exception.sqlgram"
[set theory] relational closure (relational closure related theorem)
32GB Jetson Orin SOM 不能刷机问题排查
Personal outlook | looking forward to the future from Xiaobai's self analysis and future planning
[untitled]
