# Gorila

> [Massively Parallel Methods for Deep Reinforcement Learning](https://arxiv.org/pdf/1507.04296.pdf)

我们提出了第一个用于深层强化学习的大规模分布式架构。 该体系结构使用了四个主要组件：产生新行为的并行的actor; 通过存储经验训练的并行leaner; 分布式神经网络来表示价值函数或行为策略; 一个分布式的经验池。 使用我们的架构来实施Deep Q-Network算法， 并应用于Arcade Learning Environment的49款Atari2600游戏，使用相同的超参数，在49场比赛中有41场超过了非分布式DQN，并且还减少了在大多数比赛中按照一定数量的顺序实现这些结果所需的时间。

## 方法

![](/files/-LaNIJW2Xq4nAjHT-Qki)

## 伪代码

![](/files/-LaNIJW4T_RWV5PrHA5a)


---

# Agent Instructions: Querying This Documentation

If you need additional information that is not directly available in this page, you can query the documentation dynamically by asking a question.

Perform an HTTP GET request on the current page URL with the `ask` query parameter:

```
GET https://hujian.gitbook.io/deep-reinforcement-learning/fang-fa/jie-ji-you-xi/distributed-deep-q-learning.md?ask=<question>
```

The question should be specific, self-contained, and written in natural language.
The response will contain a direct answer to the question and relevant excerpts and sources from the documentation.

Use this mechanism when the answer is not explicitly present in the current page, you need clarification or additional context, or you want to retrieve related documentation sections.
