AdaGradã¯å¦ç¿çãèªå調æ´ãã¦ãããå¾é æ³ã®äºç¨®ã§ãããããªäººã絶è³ãã¦ãã¾ãã å¾é ã足ãè¾¼ãæã«ãå次å ãã¨ã«ä»ã¾ã§ã®å¾é ã®2ä¹åãã¨ã£ã¦ããã¦ããã®å¹³æ¹æ ¹ã§å²ã£ã¦ãããã ãã¨ãæãããã·ã³ãã«ã§ãã Adaptive Subgradient Methods for Online Learning and Stochastic Optimization John Duchi, Elad Hazan, Yoram Singer. JMLR 2011. ä¸åº¦ã @echizen_tm ãããããã°ãæ¸ãã¦ã¾ããã AdaGrad+RDAãå®è£ ãã¾ããã é常ã®SGDãªã©ã¯å¦ç¿çãã ãã ãæ¸è¡°ãããªããå¾é ã足ãã¦ããããã§ãããã©ã®æ§ã«æ¸è¡°ããããã¨ããåé¡ã«ãã¤ãé ãæ©ã¾ãã¾ãã AdaGradã§ã¯æåã®å¦ç¿çããå¤ããä¸ãã¾ãããæ¸è¡°ã®ããæ¹ãæ¸è¡°çã¨ãã£ããã¤ãã¼ãã©ã¡ã¼ã¿ãã
{{#tags}}- {{label}}
{{/tags}}