硕士算法2010_第7章并行计算基础绪论.ppt

  1. 1、原创力文档(book118)网站文档一经付费(服务费),不意味着购买了该文档的版权,仅供个人/单位学习、研究之用,不得用于商业用途,未经授权,严禁复制、发行、汇编、翻译或者网络传播等,侵权必究。。
  2. 2、本站所有内容均由合作方或网友上传,本站不对文档的完整性、权威性及其观点立场正确性做任何保证或承诺!文档内容仅供研究参考,付费前请自行鉴别。如您付费,意味着您自己接受本站规则且自行承担风险,本站不退款、不进行额外附加服务;查看《如何避免下载的几个坑》。如果您已付费下载过本站文档,您可以点击 这里二次下载
  3. 3、如文档侵犯商业秘密、侵犯著作权、侵犯人身权等,请点击“版权申诉”(推荐),也可以打举报电话:400-050-0827(电话支持时间:9:00-18:30)。
查看更多
* * Shared memory. Distributed memory. Virtual shared memory * Synchronously:同步地 单指令多数据流)能够复制多个操作数,并把它们打包在大型寄存器的一组指令集。 操作数是汇编语言指令的一个字段。 * Via:通过 Explicit:明确的 * Circuitry:电路 * Synchronized:同步 * vice versa:反之亦然 Concurrent:并发 CRCW存在是因为存储中存在BANK,可以支持同时写而不会相互影响; CREW正确而CRCW不一定正确是因为系统中不一定支持互斥同步。 * 算法设计与分析 李洪伟 博士 电子科技大学计算机科学与工程学院 hongweili@ /teacher/teacher.aspx?id=298 * 第7章 并行计算基础 并行计算机 并行计算机体系结构 并行计算机存储模型 多处理器中高速缓存一致性问题 并行计算机通信机制 静态网络 动态网络 并行计算机的消息传递方式 互连网络的路由选择 并行计算模型 PRAM模型 Goals of Parallel Computing Serial computing : One processor executes a series of instructions to produce a result. Parallel computing : Produce the same result using multiple processors. In practice performance depends on the manner in which the problem is divided between the processors. Ideally want a program running on P processors to execute P times faster. Want each processor to perform a similar amount of work, i.e., ensure load balancing. * * 并行计算机体系结构变化趋势 * Classification of Parallel Architectures * SIMD Architecture Single Instruction Multiple Data. Each processor has its own memory where it keeps its data. Every processor synchronously executes same instructions on its local data. Instructions issued by controller. Processors can communicate with each other. e.g. DAP, CM200 * * MIMD Architecture Multiple Instruction Multiple Data Several independent processors capable of executing separate programs. Further subdivision on relationship between processors and memory. Shared memory. Distributed memory. Virtual shared memory * Shared Memory Small number of processors which each have access to a global memory store. Communications via write/reads to memory. Simple to program (no explicit communications). Poor scaling due to memory access bottleneck. Distributed Memory Each processor has its own local memory. Processors connected via some interconnect mechanism. Processors communicate via explicit message passing. Local memory access quicker than remote memory access

文档评论(0)

1112111 + 关注
实名认证
内容提供者

该用户很懒,什么也没介绍

1亿VIP精品文档

相关文档