1. The document discusses various statistical and neural network-based models for representing words and modeling semantics, including LSI, PLSI, LDA, word2vec, and neural network language models.
2. These models represent words based on their distributional properties and contexts using techniques like matrix factorization, probabilistic modeling, and neural networks to learn vector representations.
3. Recent models like word2vec use neural networks to learn word embeddings that capture linguistic regularities and can be used for tasks like analogy-making and machine translation.
The document summarizes a research paper that compares the performance of MLP-based models to Transformer-based models on various natural language processing and computer vision tasks. The key points are:
1. Gated MLP (gMLP) architectures can achieve performance comparable to Transformers on most tasks, demonstrating that attention mechanisms may not be strictly necessary.
2. However, attention still provides benefits for some NLP tasks, as models combining gMLP and attention outperformed pure gMLP models on certain benchmarks.
3. For computer vision, gMLP achieved results close to Vision Transformers and CNNs on image classification, indicating gMLP can match their data efficiency.
1) The document discusses the development history and planned features of Chainer, a deep learning framework.
2) It describes Chainer's transition to a new model structure using Links and Chains to define networks in a more modular and reusable way.
3) The new structure will allow for easier saving, loading, and composition of network definitions compared to the previous FunctionSet/Optimizer approach.
Literate Computing for Infrastructure - インフラ・コード化の実践におけるIPython (Jupyter) Not...No Bu
Presented at SC2015-6 on 6/3/2015 for ..
Infrastructure as Code meets IPython Notebook to be Literate Computing
IEICE Tech. Rep., vol. 115, no. 72, SC2015-6, pp. 27-32, June 2015.
Abstract: Cloud has put the pressure to rapidly build systems and frequently re-configure services, then Infrastructure as Code has come beyond the simple automation. The approach treats the configuration of systems the same way that software source code is treated. Infrastructure is validated and processed “as Code” with management tools. However, as Code is not limited only about the intelligent automation, but also about the communication based on code for reviewing, reproducing, customizing, and reusing. It is as important to be able to share information and processes with others, as to actually automate complex operations for infrastructures. IPython Notebook is a useful tool to both describe automated operations with code (and configuration data) and share predicted and reproducible outcomes with others, technical and non-technical alike.
IPython Notebook is a “literate computing” tool, which enables us to share stories about infrastructure’s design and elaborated workflows. We will share our experience how the literate stories are also useful for various customer communications as tracing individual issue, promoting self-administration etc.
Keywords DevOps, Infrastructure as Code, Literate Computing, IPython Notebook, Jupyter
インフラ・コード化の実践におけるIPython Notebookの適用
信学技報, vol. 115, no. 72, SC2015-6, pp. 27-32, 2015年6月
あらまし: クラウドサービスの浸透により,サービスの構築・再構築の機会が増加するのに伴って,作業手順をすべてCodeとして記述するInfrastructures as Codeというアプローチが着目されている.ここでの“as Code”は作業手順の正当性がプログラムコードのように,また実行結果も機械的に検証可能であるという意味合いで捉えられがちであるが,むしろ個々の作業の再現性を保証し,その上で作業をカスタマイズ・再利用すると言ったプロセス自体を,Codeとして見える化し,伝達可能にすることにこそ意義がある.DevOpsに於いては,何某かを実際に構築したり機械化したりするだけではなく,設計情報,運用状態を伝達・共有できるようにすることが重要である.
“Literate Computing”ツールと呼ばれ,ワークフローと実行結果を一体としてドキュメント化できるIPython Notebookを,基盤の構築,運用に適用する方式を提案すると共に,具体的な適用によってワークフローをどのように改善することができたかを報告する.
キーワード DevOps, Infrastructure as Code, Literate Computing, IPython Notebook, Jupyter
https://djangocongress.jp/#talk-10
OpenTelemetryは、複数のプロセス、システムをまたがってアプリケーションの処理を追跡する分散トレースの仕組みを提供するフレームワークで、2021年春に1.0.0がリリースされました。このライブラリを活用し、Djangoアプリおよび周辺システムの処理を追跡する方法について紹介します。
Google Slide(スライド内のリンクをクリックできます)
https://docs.google.com/presentation/d/e/2PACX-1vRtqRQ6USDeV32_aTPjSaNXpKdn5cbitkmiX9ZfgwXVE-mh74I4eICFOB8rWGz0LPUIEfXn3APRKcrU/pub
コード
https://github.com/shimizukawa/try-otel/tree/20221112-djangocongressjp2022
Let's trace web system processes with opentelemetry djangocongress jp 2022
Preferred Networks is a Japanese AI startup founded in 2014 that develops deep learning technologies. They presented at CEATEC JAPAN 2018 on their research using convolutional neural networks for computer vision tasks like object detection. They discussed techniques like residual learning and how they have achieved state-of-the-art results on datasets like COCO by training networks on large amounts of data using hundreds of GPUs.
Preferred Networks was founded in 2008 and has focused on deep learning research, developing the Chainer and CuPy frameworks. It has applied its technologies to areas including computer vision, natural language processing, and robotics. The company aims to build AI that is helpful, harmless, and honest through techniques like constitutional AI that help ensure systems behave ethically and avoid potential issues like bias, privacy concerns, and loss of control.
Preferred Networks was founded in 2008 and has developed technologies such as Chainer and CuPy. It focuses on neural networks, natural language processing, computer vision, and GPU computing. The company aims to build general-purpose AI through machine learning and has over 500 employees located in Tokyo and San Francisco.
This document discusses Preferred Networks' open source activities over the past year. It notes that Preferred Networks published 10 blog posts and tech talks on open source topics and uploaded 3 videos to their Youtube channel. It also mentions growing their open source community to over 120 members and contributors across 3 major open source projects. The document concludes by reaffirming Preferred Networks' commitment to open source software, blogging, and tech talks going forward.
1. This document discusses the history and recent developments in natural language processing and deep learning. It covers seminal NLP papers from the 1990s through 2000s and the rise of neural network approaches for NLP from 2003 onward.
2. Recent years have seen increased research and investment in deep learning, with many large companies establishing AI labs in 2012-2014 to focus on neural network techniques.
3. The document outlines some popular deep learning architectures for NLP tasks, including neural language models, word2vec, sequence-to-sequence learning, and memory networks. It also introduces the Chainer deep learning framework for Python.
1. The document discusses knowledge representation and deep learning techniques for knowledge graphs, including embedding models like TransE, TransH, and neural network models.
2. It provides an overview of methods for tasks like link prediction, question answering, and language modeling using recurrent neural networks and memory networks.
3. The document references several papers on knowledge graph embedding models and their applications to natural language processing tasks.
This document provides an overview of preferred natural language processing infrastructure and techniques. It discusses recurrent neural networks, statistical machine translation tools like GIZA++ and Moses, voice recognition systems from NICT and NTT, topic modeling using latent Dirichlet allocation, dependency parsing with minimum spanning trees, and recursive neural networks for natural language tasks. References are provided for several papers on these methods.
1. The document discusses the history and recent developments in natural language processing and deep learning. It provides an overview of seminal NLP papers from the 1990s to 2010s and deep learning architectures from 2003 to present.
2. Key deep learning models discussed include neural language models, word2vec, convolutional neural networks, and LSTMs. The document also notes the increasing interest and research in deep learning starting in 2012 by tech companies like Google, Facebook and Baidu.
3. Application examples mentioned include search engines, conversational agents, social media and news summarization tools.
EMNLP2014読み会 "Efficient Non-parametric Estimation of Multiple Embeddings per ...Yuya Unno
1. The document presents the Multi Sense Skip-gram (MSSG) model for learning multiple embeddings per word in vector space.
2. MSSG assigns a separate embedding to each sense of a word using a context vector. It extends the Skip-gram model by learning sense-specific embeddings.
3. The Non-Parametric MSSG (NP-MSSG) model extends MSSG by using a non-parametric approach to learn the context vectors instead of fixed vectors, allowing an unbounded number of senses per word.
20. 擬似コードで⽐比較する
define-and-run
# 構築
x = Variable(‘x’)
y = Variable(‘y’)
z = x + 2 * y
# 評価
for xi, yi in data:
eval(z, x=xi, y=yi))
define-by-run
# 構築と評価が同時
for xi, yi in data:
x = Variable(xi)
y = Variable(yi)
z = x + 2 * y
20
データを⾒見見ながら
違う処理理をしてもよい
21. 計算グラフで⽐比較する
21
s = 0
for x in [1, 2, 3]:
s += x
s
x
+
x
+
x
+ ss
x
+ s
define-and-runで
ループを作る
define-by-runでは
すべて展開される
33. CuPyとNumPyの⽐比較
import numpy
x = numpy.array([1,2,3], numpy.float32)
y = x * x
s = numpy.sum(y)
print(s)
import cupy
x = cupy.array([1,2,3], cupy.float32)
y = x * x
s = cupy.sum(y)
print(s)
33
34. CuPyはどのくらい早いの?
l 状況しだいですが、最⼤大数⼗十倍程度度速くなります
def test(xp):
a = xp.arange(1000000).reshape(1000, -1)
return a.T * 2
test(numpy)
t1 = datetime.datetime.now()
for i in range(1000):
test(numpy)
t2 = datetime.datetime.now()
print(t2 -t1)
test(cupy)
t1 = datetime.datetime.now()
for i in range(1000):
test(cupy)
t2 = datetime.datetime.now()
print(t2 -t1)
34
時間
[ms]
倍率率率
NumPy 2929 1.0
CuPy 585 5.0
CuPy +
Memory Pool
123 23.8
Intel Core i7-4790 @3.60GHz,
32GB, GeForce GTX 970