Building PetaByte Servers.ppt
《Building PetaByte Servers.ppt》由会员分享,可在线阅读,更多相关《Building PetaByte Servers.ppt(63页珍藏版)》请在麦多课文档分享上搜索。
1、1,Building PetaByte Servers,Jim Gray Microsoft Research GrayM http:/www.Research.M 103 Mega 106 Giga 109 Tera 1012 today, we are here Peta 1015 Exa 1018,2,Outline,The challenge: Building GIANT data stores for example, the EOS/DIS 15 PB system Conclusion 1 Think about MOX and SCANS Conclusion 2: Thin
2、k about Clusters,3,The Challenge - EOS/DIS,Antarctica is melting - 77% of fresh water liberated sea level rises 70 meters Chico & Memphis are beach-front property New York, Washington, SF, LA, London, Paris Lets study it! Mission to Planet Earth EOS: Earth Observing System (17B$ = 10B$) 50 instrumen
3、ts on 10 satellites 1997-2001 Landsat (added later) EOS DIS: Data Information System: 3-5 MB/s raw, 30-50 MB/s processed. 4 TB/day, 15 PB by year 2007,4,The Process Flow,Data arrives and is pre-processed. instrument data is calibrated, gridded averaged Geophysical data is derived Users ask for store
4、d data OR to analyze and combine data. Can make the pull-push split dynamically,Pull Processing,Push Processing,Other Data,5,Designing EOS/DIS,Expect that millions will use the system (online) Three user categories: NASA 500 - funded by NASA to do science Global Change 10 k - other dirt bags Interne
5、t 20 m - everyone elseGrain speculatorsEnvironmental Impact ReportsNew applications = discovery & access must be automatic Allow anyone to set up a peer- node (DAAC & SCF) Design for Ad Hoc queries, Not Standard Data Products If push is 90%, then 10% of data is read (on average). = A failure: no one
6、 uses the data, in DSS, push is 1% or less.= computation demand is enormous (pull:push is 100: 1),6,The architecture,2+N data center design Scaleable OR-DBMS Emphasize Pull vs Push processing Storage hierarchy Data Pump Just in time acquisition,7,Obvious Point: EOS/DIS will be a cluster of SMPs,It n
7、eeds 16 PB storage = 1 M disks in current technology = 500K tapes in current technology It needs 100 TeraOps of processing = 100K processors (current technology) and 100 Terabytes of DRAM 1997 requirements are 1000x smaller smaller data rate almost no re-processing work,8,2+N data center design,dupl
8、ex the archive (for fault tolerance) let anyone build an extract (the +N) Partition data by time and by space (store 2 or 4 ways). Each partition is a free-standing OR-DBBMS (similar to Tandem, Teradata designs). Clients and Partitions interact via standard protocols OLE-DB, DCOM/CORBA, HTTP,9,Hardw
9、are Architecture,2 Huge Data Centers Each has 50 to 1,000 nodes in a cluster Each node has about 25250 TB of storage SMP .5Bips to 50 Bips 20K$ DRAM 50GB to 1 TB 50K$ 100 disks 2.3 TB to 230 TB 200K$ 10 tape robots 25 TB to 250 TB 200K$ 2 Interconnects 1GBps to 100 GBps 20K$ Node costs 500K$ Data Ce
10、nter costs 25M$ (capital cost),10,Scaleable OR-DBMS,Adopt cluster approach (Tandem, Teradata, VMScluster,) System must scale to many processors, disks, links OR DBMS based on standard object model CORBA or DCOM (not vendor specific) Grow by adding components System must be self-managing,11,Storage H
11、ierarchy,Cache hot 10% (1.5 PB) on disk. Keep cold 90% on near-line tape. Remember recent results on speculation (more on this later MOX/GOX/SCANS),12,Data Pump,Some queries require reading ALL the data (for reprocessing) Each Data Center scans the data every 2 weeks. Data rate 10 PB/day = 10 TB/nod
12、e/day = 120 MB/s Compute on demand small jobsless than 1,000 tape mountsless than 100 M disk accessesless than 100 TeraOps.(less than 30 minute response time) For BIG JOBS scan entire 15PB database Queries (and extracts) “snoop” this data pump.,13,Just-in-time acquisition 30%,Hardware prices decline
13、 20%-40%/year So buy at last moment Buy best product that day: commodity Depreciate over 3 years so that facility is fresh. (after 3 years, cost is 23% of original). 60% decline peaks at 10M$,1996,EOS DIS Disk Storage Size and Cost,1994,1998,2000,2002,2004,2006,2008,Storage Cost M$,Data Need TB,assu
14、me 40% price decline/year,14,Problems,HSM Design and Meta-data Ingest Data discovery, search, and analysis reorg-reprocess disaster recovery cost,15,What this system teaches us,Traditional storage metrics KOX: KB objects accessed per second $/GB: Storage cost New metrics: MOX: megabyte objects acces
15、sed per second SCANS: Time to scan the archive,16,Thesis: Performance =Storage Accesses not Instructions Executed,In the “old days” we counted instructions and IOs Now we count memory references Processors wait most of the time,17,The Pico Processor,1 M SPECmarks106 clocks/fault to bulk ramEvent-hor
16、izon on chip.VM reincarnatedMulti-program cache,Terror Bytes!,18,Storage Latency: How Far Away is the Data?,Registers,On Chip Cache,On Board Cache,Memory,Disk,1,2,10,100,Tape /Optical,Robot,10,9,10,6,Sacramento,This Campus,This Room,My Head,10 min,1.5 hr,2 Years,1 min,Pluto,2,000 Years,Andromeda,19,
17、DataFlow Programming Prefetch & Postwrite Hide Latency,Cant wait for the data to arrive (2,000 years!) Need a memory that gets the data in advance ( 100MB/S) Solution: Pipeline data to/from the processor Pipe data from source (tape, disc, ram.) to cpu cache,20,MetaMessage: Technology Ratios Are Impo
18、rtant,If everything gets faster&cheaper at the same rate THEN nothing really changes. Things getting MUCH BETTER: communication speed & cost 1,000x processor speed & cost 100x storage size & cost 100x Things staying about the same speed of light (more or less constant) people (10x more expensive) st
19、orage speed (only 10x better),21,Trends: Storage Got Cheaper,$/byte got 104 better $/access got 103 better capacity grew 103 Latency improved 10 Bandwidth improved 10,Year,1960,1970,1980,1990,2000,Disk (kB),Storage Capacity,RAM (b),Tape (kB),Unit Storage Size,22,Trends: Access Times Improved Little,
20、1,1960,1970,1980,1990,2000,Processors,Year,Instructions / second,Processor Speedups,Bits / second,WANs,1e 2,1e 1,1e 0,1e -1,1e-2,1e-3,1e-4,1e-5,1e-6,1e-7,1e 3,1960,1970,1980,1990,2000,Tape,Disk,RAM,Year,Access Times Improved Little,23,Trends: Storage Bandwidth Improved Little,1960,1970,1980,1990,200
21、0,Tape,Disk,RAM,Year,Transfer Rates Improved Little,1e -1,1,1960,1970,1980,1990,2000,Processors,Year,Processor Speedups,WANs,24,Todays Storage Hierarchy : Speed & Capacity vs Cost Tradeoffs,Typical System (bytes),Size vs Speed,Access Time (seconds),10,-9,10,-6,10,-3,10,0,10,3,Cache,Main,Secondary,Di
- 1.请仔细阅读文档,确保文档完整性,对于不预览、不比对内容而直接下载带来的问题本站不予受理。
- 2.下载的文档,不会出现我们的网址水印。
- 3、该文档所得收入(下载+内容+预览)归上传者、原创作者;如果您是本文档原作者,请点此认领!既往收益都归您。
下载文档到电脑,查找使用更方便
2000 积分 0人已下载
下载 | 加入VIP,交流精品资源 |
- 配套讲稿:
如PPT文件的首页显示word图标,表示该PPT已包含配套word讲稿。双击word图标可打开word文档。
- 特殊限制:
部分文档作品中含有的国旗、国徽等图片,仅作为作品整体效果示例展示,禁止商用。设计者仅对作品中独创性部分享有著作权。
- 关 键 词:
- BUILDINGPETABYTESERVERSPPT
