当前位置: 首页 > news >正文

商标 做网站 是几类群晖的网站开发

商标 做网站 是几类,群晖的网站开发,游戏推广员到底犯不犯法,西安手机网站建站论文原文#xff1a;[1810.00826] How Powerful are Graph Neural Networks? (arxiv.org) 英文是纯手打的#xff01;论文原文的summarizing and paraphrasing。可能会出现难以避免的拼写错误和语法错误#xff0c;若有发现欢迎评论指正#xff01;文章偏向于笔记#x… 论文原文[1810.00826] How Powerful are Graph Neural Networks? (arxiv.org) 英文是纯手打的论文原文的summarizing and paraphrasing。可能会出现难以避免的拼写错误和语法错误若有发现欢迎评论指正文章偏向于笔记谨慎食用 1. 省流版 1.1. 心得 ①Emm, 数学上的解释性确实很强了 ②他一直在...在说引理 1.2. 论文框架图 2. 论文逐段精读 2.1. Abstract ①Even though the occurrence of Graph Neural Networks (GNNs) changes graph representation learning to a large extent, it and its variants are all limited in representation abilities. 2.2. Introduction ①Briefly introduce how GNN works (combining node information from k-hop neighbors and then pooling) ②The authors hold the view that ⭐ other graph models mostly based on plenty experimental trial-and-errors rather than theoretical understanding ③They combine GNNs and the Weisfeiler-Lehman (WL) graph isomorphism test to build a new framework, which relys on multisets ④GIN is excellent in distinguish, capturing and representaion heuristics  n.[U] (formal) 探索法启发式 heuristic  adj.(教学或教育)启发式的 2.3. Preliminaries 1Their definition ①They define two tasks: node classicifation with node label  and graph classification with graph label  2Other models ①The authors display the function of GNN in the -th layer: where only  is initialized to  其余细节就不多说了在GNN的笔记里都有 ②Pooling layer of GraphSAGE, the AGGREGATE function is: where MAX is element-wise max-pooling operator; is learnable weight matrix; and followed by concatenated COMBINE and linear mapping  ③AGGREGATE and COMBINE areintegrated in GCN: ④Lastly follows a READOUT layer to get final prediction answer: where the READOUT function can be different forms 3Weisfeiler-Lehman (WL) test ①WL firstly aggregates nodes and their neighborhoods and then hashs the labels (??hash?这好吗) ②Based on WL, WL subtree kernel was proposed to evaluate the similarity between graphs ③A subtree of height s root node is the node at -th iteration permutation  n.置换;排列(方式);组合(方式) 2.4. Theoretical framework: overview ①The framework overview ②Multiset: is a 2-tuple , where where  is the underlying set of  that is formed from its distinct elements, and  gives the multiplicity of the elements 我没有太懂这句话欸 ③They are not allowed that GNN map different neighbors to the same representation. Thus, the aggregation must be injective 我也不造为啥 2.5. Building powerful graph neural networks ①They define Lemma 2, namely WL graph isomorphism test is able to correctly distinguish non-isomorphic graphs ②Theorem 3 完全没看懂 ③Lemma 4: If input feature space is countable, then the space of node hidden features  is also countable 2.5.1. Graph isomorphism network (GIN) ①Lemma 5: there is  , which makes  unique in  . Also there is  ②Corollary 6: there is unique  and . ③Finally, the update function of GIN can be: 2.5.2. Graph-level readout of GIN ①Sum, mean and max aggregators: ②The fail examples when the different  and  map the same embedding: where (a) represents all the nodes are the same, only sum can distinguish them; blue in (b) represents the max, thus max fails to distinguish as well; same in (c). 盲猜这里其实蓝色v自己是一个节点但是没有考虑自己的特征而是纯看1-hop neighborhoods ③They change the READOUT layer to: 2.6. Less powerful but still interesting GNNs They designed ablation studies 2.6.1. 1-layer perceptrons are not sufficient ①1-layer perceptrons are akin to linear mapping, which is far insufficient for distinguishing ②Lemma 7: notwithstanding multiset  is different from , they might get the same results:  2.6.2. Structures that confuse mean and max-pooling 这一节的内容在2.5.2.②的图下已经解释过了 2.6.3. Mean learns distributions ①Collary 8: there is a function . If and only if multisets  and  are the same distribution,  ②When statistical and distributional information in graph cover more important part, mean aggregator performs better. But when structure is valued more, mean aggregator may do worse. ③Sum and mean aggregator may be similar when node features are multifarious and hardly repeat 2.6.4. Max-pooling learns sets with distinct elements ①Max aggregator focus on learning the structure of graph (原文用的skeleton而不是structure), and it has a certain ability to resist noise and outliers ②For max function , if and only if  and  have the same underlying set,  2.6.5. Remarks on other aggregators ①They do not cover the analysis of weighted average via attention or LSTM pooling 2.7. Other related work ①Traditional GNN does not provide enough math explanation ②Exceptionally, RKHS of graph kernels (?) is able to approximate measurable functions in probability ③Also, they can hardly generalize to multple architectures 2.8. Experiments 1Datasets ①Dataset: 9 graph classification benchmarks: 4 bioinformatics datasets (MUTAG, PTC, NCI1, PROTEINS) and 5 social network datasets (COLLAB, IMDB-BINARY, IMDB-MULTI, REDDITBINARY and REDDIT-MULTI5K) ②Social networks are lack of node features, then they set node vectors as the same in REDDIT and use one hot encoding for others 2Mondels and configurations ①They set two variants, the one is GIN-ε, which adopts gradient descent, the other one is GIN-0, which is a little bit simpler. ②Performances of different variants on different datasets ③Validation: 10-fold LIB-SVM ④Layers: 5, includes input layer, and each MLP takes two layers ⑤Normalization: batch normalization for all hiden layers ⑥Optimizer: Adam ⑦Learning rate: 0.01 at first and substract 0.5/50 epochs ⑧Number of hidden units, hyper parameter: 16 or 32 ⑨Batch size: 32 or 128 ⑩Drop out ratio: 0 or 0.5 ⑪Epoch: the best one in 10-fold 3Baselines ①WL subtree kernel ②Diffusionconvolutional neural networks (DCNN), PATCHY-SAN (Niepert) and Deep Graph CNN (DGCNN) ③Anonymous Walk Embeddings (AWL) 2.8.1. Results 1Training set performance ①Training set accuracy figure was showed above ②WL always performs better than GNN due to its strong classifying ability. However, WL can not present the node features combination, which may limit in the future 2Test set performance ①Test set classification accuracies ②GIN-0 obviously outperforms others 2.9. Conclusion They give theoretical foundations of graph structure and discuss the performances of variants of GNN. Then, they designed a strong GNN, named GIN to achieve more accurate classification. Furthermore, they think researching the generalization for GNNs is also promising. 3. Reference List Xu, K. et al. (2019) How Powerful are Graph Neural Networks?, ICLR 2019. doi: https://doi.org/10.48550/arXiv.1810.00826
http://wiki.neutronadmin.com/news/165191/

相关文章:

  • 女生做网站编辑怎么样四川建筑人才招聘网
  • 博罗东莞网站建设全网关键词优化公司哪家好
  • 做网站的地方wordpress点击分类目录空白
  • 个人域名能做网站吗视频拍摄团队
  • 深圳做营销网站公司简介网页游戏网站哪个好
  • 做一个静态网站需要多少钱一次备案多个网站
  • 自己做网站步骤 域名ui设计就业前景
  • 网站建设调研问卷怎么做免费网站
  • 网页制作与设计实训目的怎么写面试问你如何快速优化网站
  • 做网站是干嘛直播系统平台搭建
  • 篮球运动装备网站模板网页app开发培训班
  • 花生壳域名做网站怎么在微信上做公众号
  • 打电话推销好还是做网站推广好东莞东城社保局电话
  • 罗湖商城网站建设哪家服务周到培训如何优化网站
  • 龙凤网站建设云聚达宣传片拍摄公司排名
  • 别样网站建设成为架构师有多难
  • 网站备案登陆用户名是什么小型网站建设公司价格
  • 东城企业网站开发网站主色调简介怎么说
  • 合肥建设局网站领导宜州市住房保障和城乡建设局网站
  • 如何在网站上做免费广告WordPress文章按钮
  • 规划设计网站推荐网站建设 的公司
  • 网站用户注册页面怎么做弹窗广告最多的网站
  • 秦皇岛住房建设网站企业微信app下载安装电脑版
  • 织梦模板大气网站建设类网站模板下载wordpress推广网站插件
  • 给自己公司做个网站成都seo优化
  • mvc5 网站开发之學 pdf行业网站建设分析
  • 棋牌网站制作价格简单asp网站源码
  • 用闲置的安卓手机做网站服务器如何做测评视频网站
  • 产品展示网站含后台网站模板下载兰州建设工程信息网站
  • 安阳网站建设开发建筑模拟2015