免费注册 查看新帖 |

Chinaunix

  平台 论坛 博客 文库
最近访问板块 发新帖
楼主: yftty
打印 上一主题 下一主题

浩存 - 面向数据库,虚拟机等海量数据可同时提供NFS/iSCSI访问的集群存储系统 [复制链接]

论坛徽章:
0
81 [报告]
发表于 2005-06-01 13:42 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

看完这个帖子,第一印象不是这个项目或者牵涉的技术,而是yftty这个家伙
很能侃。

我不懂,但是我想看看,,,我该通过何种方式来旁观这个项目呢?

论坛徽章:
0
82 [报告]
发表于 2005-06-03 09:26 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

不错的说~

论坛徽章:
0
83 [报告]
发表于 2005-06-07 09:41 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

here is the current sanity testing & results

[yf@yftty xxxfs]$ tests/xxxfs_sanity -v
000010:000001:1118108292.377965:4560socket.c:63xxfs_net_connect()) Process entered
config finished, ready to do the sanity testing !
xxxFS file creation testing succeeded !
xxxFS file read testing succeeded !
xxxFS file deletion testing succeeded !
xxxFS Sanity testing pid (4560) succeeded 1 !
[yf@yftty xxxfs]$

论坛徽章:
0
84 [报告]
发表于 2005-06-07 11:38 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

项目到现在已快要过两个季度,经过这些时间的实践和思考

我的浅见是这个项目从流程来说上面的发贴所谈的已经比较完善了
从分工和组织来说,大家看下面的是否合适?

        __________           __________
       | 理论指导 |   <->;   | 开发指导 |
        ----------           ----------
         |        \         /       |
         |         \       /        |
       ------       ------       --------
      | 研发 | <->; | 开发 | <->; |  测试  |
       ------       ------       --------

另:

这样还是有问题, 晕.

论坛徽章:
0
85 [报告]
发表于 2005-06-08 09:30 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

Dan Stromberg wrote:
>; The lecturer at the recent NG storage talk at Usenix in Anaheim,
>; indicated that it was best to avoid "active/active" and get
>; "active/passive" instead.
>;
>; Does anyone:
>;
>; 1) Know what these things mean?

In the clustering world, active/active means 2 or more servers are
active at a time, either operating on separate data (and thus acting as
passive failover partners to each other), or operating on the same data
(which requires the use of a cluster filesystem or other similar
mechanism to allow coherent simultaneous access to the data).

>; 2) Know why active/passive might be preferred over active/active?

Well, if you're talking about active/passive vs. active/active with a
cluster filesystem or such, the active/passive is tons easier to
implement and get right. Plus, depending on your application, the added
complexity of a cluster filesystem might not actually buy you much more
than you could get with, say, NFS or Samba (CIFS).

--
Paul

论坛徽章:
0
86 [报告]
发表于 2005-06-08 11:00 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

http://tech.blogchina.com/53/2005-06-07/372338.html

想了解Google的企业文化,需要从Google创立时的一个插曲开始:当谢尔盖·布林(Sergey Brin)和拉里·佩奇(Larry Page)想将自己的网络梦想付诸实际,最大的障碍是,他们并没有足够的资金来购买价格昂贵的设备。于是两人花费数百美元购买了一些个人电脑来代替那些数百万美元的服务器。

在实际应用中,这些普通电脑的故障率自然要高于专业服务器。他们需要确保任何一台普通电脑发生故障时都不会影响到用户正常得出搜索结果,于是Google 决定自己开发软件工具来解决这些问题。比如Google文件系统。这种文件系统不仅能够高效处理大型数据,还能够随时应付突然发生的存储故障。配合 Google的三重备份体制,这些个人电脑组成的系统就可以完成那些服务器的工作。

而这种遇到任何问题都全力解决之的理念,极大的影响了后来Google的文化。至今,Google依旧保持着网络公司的风貌。拥有2700名员工的公司总部里有900人是技术人员,而且在这里没有几间办公室。在施密特衣柜般的小办公室楼下,布林和佩奇共用一间办公室。而那里就像一间大学宿舍,里面摆着冰球装备、滑板和遥控飞机模型、懒人椅等等。

...

没有人质疑Google拥有魔幻般的技术和创新,但没有一家伟大的公司仅仅依靠出色的技术而成为世界级的公司。伟大的公司需要伟大的管理来帮助公司更上层楼。谁是Google的灵魂?当然是布林、佩奇再加上施密特组成的三人组。但谈到管理层面,49岁的施密特的确起到了至关重要的作用。

49岁的施密特曾经是Sun公司的CTO以及Novell公司的CEO,他至今仍清晰记得刚到这家公司时董事会对他的交待:“别把公司弄糟了,艾利克。公司的起点非常非常好,可别进行太大的改革。”他完全理解投资者的担心,他们不想这家创造力十足的公司变得僵化死板。

1999年施密特刚到这家公司的时候这里根本谈不上有什么管理,但他也不想照搬传统大公司那一套管理方法,他希望根据实际情况形成Google自己的管理模式。大多数情况下施密特和2位创始人一起行动,作出决策。通常情况下是施密特主持管理层会议,而2位创始人主持员工会议。当遇到重大问题需要解决的时候,Google3人组就会根据少数服从多数的基本规则作出决定。并且许多决定他们是当着员工的面得出结果的。公司管理层刻意保持企业文化中率直、自由的工程师文化,他们认为这是他们抗衡Yahoo和微软这样大规模公司的有力武器。

哈佛商学院教授大卫·友菲(David Yoffie)却并不看好这种管理模式:“如果很多人同时作决定,那等于没有决定任何事情。在Google每天会同时作出成千上万的计划,需要有一个人作出最终决断。”

施密特表示实际上他所扮演的角色更倾向于COO。他以雅虎和eBay举例来说,在这些公司里都是创始人来制定远景战略,尽管他们并不拥有首席执行官的头衔。但施密特的支持者认为,这名CEO的个人风格掩盖了他在公司中的实际地位。而曾经担任CEO的佩奇如今担任产品总裁。前董事长布林则担任技术总裁。而施密特则在过去的4年中为Google搭建了完善的架构。

布林和佩奇的管理哲学完全源于他们当初所在的斯坦福大学计算机科学实验室。Google的经理很少要求那些工程师去完成什么项目,取而代之的则是公司会宣布一个100项优先完成项目列表,工程师们根据自己的喜好参加不同的流动工作组,以周或者月为时间单位完成工作。

论坛徽章:
0
87 [报告]
发表于 2005-06-10 23:49 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

openMosix:
5.1 What Is openMosix?
Basically, the openMosix software includes both a set of kernel patches and support tools. The patches extend the kernel to provide support for moving processes among machines in the cluster. Typically, process migration is totally transparent to the user. However, by using the tools provided with openMosix, as well as third-party tools, you can control the migration of processes among machines.

Let's look at how openMosix might be used to speed up a set of computationally expensive tasks. Suppose, for example, you have a dozen files to compress using a CPU-intensive program on a machine that isn't part of an openMosix cluster. You could compress each file one at a time, waiting for one to finish before starting the next. Or you could run all the compressions simultaneously by starting each compression in a separate window or by running each compression in the background (ending each command line with an &. Of course, either way will take about the same amount of time and will load down your computer while the programs are running.

However, if your computer is part of an openMosix cluster, here's what will happen: First, you will start all of the processes running on your computer. With an openMosix cluster, after a few seconds, processes will start to migrate from your heavily loaded computer to other idle or less loaded computers in the clusters. (As explained later, because some jobs may finish quickly, it can be counterproductive to migrate too quickly.) If you have a dozen idle machines in the cluster, each compression should run on a different machine. Your machine will have only one compression running on it (along with a little added overhead) so you still may be able to use it. And the dozen compressions will take only a little longer than it would normally take to do a single compression.

If you don't have a dozen computers, or some of your computers are slower than others, or some are otherwise loaded, openMosix will move the jobs around as best it can to balance the load. Once the cluster is set up, this is all done transparently by the system. Normally, you just start your jobs. openMosix does the rest. On the other hand, if you want to control the migration of jobs from one computer to the next, openMosix supplies you with the tools to do just that.

OSCAR:

Setting up a cluster can involve the installation and configuration of a lot of software as well as reconfiguration of the system and previously installed software. OSCAR (Open Source Cluster Application Resources) is a software package that is designed to simplify cluster installation. A collection of open source cluster software, OSCAR includes everything that you are likely to need for a dedicated, high-performance cluster. OSCAR takes you completely through the installation of your cluster. If you download, install, and run OSCAR, you will have a completely functioning cluster when you are done.

The design goals for OSCAR include using the best-of-class software, eliminating the downloading, installation, and configuration of individual components, and moving toward the standardization of clusters. OSCAR, it is said, reduces the need for expertise in setting up a cluster. In practice, it might be more fitting to say that OSCAR delays the need for expertise and allows you to create a fully functional cluster before mastering all the skills you will eventually need. In the long run, you will want to master those packages in OSCAR that you come to rely on. OSCAR makes it very easy to experiment with packages and dramatically lowers the barrier to getting started.

OSCAR was created and is maintained by the Open Cluster Group (http://www.openclustergroup.org), an informal group dedicated to simplifying the installation and use of clusters and broadening their use. Over the years, a number of organizations and companies have supported the Open Cluster Group, including Dell, IBM, Intel, NCSA, and ORNL, to mention only a few.

Because OSCAR is an extensive collection of software, it is beyond the scope of this book to cover every package in detail. Most of the software in OSCAR is available as standalone versions, and many of the key packages included by OSCAR are described in later chapters in this book. Consequently, this chapter focuses on setting up OSCAR and on software unique to OSCAR. By the time you have finished this chapter, you should be able to judge whether OSCAR is appropriate for your needs and know how to get started.

Rocks:
NPACI Rocks is a collection of open source software for building a high-performance cluster. The primary design goal for Rocks is to make cluster installation as easy as possible. Unquestionably, they have gone a long way toward meeting this goal. To accomplish this, the default installation makes a number of reasonable assumptions about what software should be included and how the cluster should be configured. Nonetheless, with a little more work, it is possible to customize many aspects of Rocks.

When you install Rocks, you will install both the clustering software and a current version of Red Hat Linux updated to include security patches. The Rocks installation will correctly configure various services, so this is one less thing to worry about. Installing Rocks installs Red Hat Linux, so you won't be able to add Rocks to an existing server or use it with some other Linux distribution.

Default installations tend to go very quickly and very smoothly. In fact, Rocks' management strategy assumes that you will deal with software problems on a node by reinstalling the system on that node rather than trying to diagnose and fix the problem. Depending on hardware, it may be possible to reinstall a node in under 10 minutes. Even if your systems take longer, after you start the reinstall, everything is automatic, so you don't need to hang around.

In this chapter, we'll look briefly at how to build and use a Rocks cluster. This coverage should provide you with enough information to decide whether Rocks is right for you. If you decide to install Rocks, be sure you download and read the current documentation. You might also want to visit Steven Baum's site。

论坛徽章:
0
88 [报告]
发表于 2005-06-14 15:49 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

This is a short taxonomy of the kinds of distributed filesystems you can find today (Febrary 2004). This was assembled with some help from Garth Gibson and Larry Jones.

Distributed filesystem - the generic term for a client/server or "network" filesystem where the data isn't locally attached to a host. There are lots of different kinds of distributed filesystems, the first ones coming out of research in the 1980s. NFS and CIFS are the most common distributed filesystems today

Global filesystem - this refers to the namespace, so that all files have the same name and path name when viewed from all hosts. This obviously makes it easy to share data across machines and users in different parts of the organization. For example, the WWW is a global namespace because a URL works everywhere. But, filesystems don't always have that property because your share definitions may not match mine, we may not see the same file servers or the same portions of those file servers.

AFS was an early provider of a global namespace - all files were organized under /afs/cellname/... and you could assemble AFS cells even from different organizations (e.g., different universities) into one shared filesystem. The Panasas filesystem (PanFS) supports a similar structure, if desired.

SAN filesystem - these provide a way for hosts to share Fibre Channel storage, which is traditionally carved into private chunks bound to different hosts. To provide sharing, a block-level metadata manager controls access to different SAN devices. A SAN Filesystem mounts storage natively in only one node, but connects all nodes to that storage and distributes block addresses to other nodes. Scalability is often an issue because blocks are a low-level way to share data placing a big burden on the metadata managers and requiring large network transactions in order to access data.

Examples include SGI cXFS, IBM GPFS, Red Hat Sistina, IBM SanFS, EMC Highroad and others.

Symmetric filesystems - A symmetric filesystem is one in which the clients also run the metadata manager code; that is, all nodes understand the disk structures. A concern with these systems is the burden that metadata management places on the client node, serving both itself and other nodes, which may impact the ability of the client to perform its intended compute jobs. Examples include Sistina GFS, GPFS, Compaq CFS, Veritas CFS, Polyserve Matrix

Asymmetric filesystems - An asymmetric filesystem is one in which there are one or more dedicated metadata managers that maintain the filesystem and its associated disk structures. Examples include Panasas ActiveScale, IBM SanFS, and Lustre. Traditional client/server filesystems like NFS and CIFS are also asymmetric.

Cluster filesystem - a distributed filesystem that is not a single server with a set of clients, but instead a cluster of servers that all work together to provide high performance service to their clients. To the clients the cluster is transparent - it is just "the filesystem", but the filesystem software deals with distributing requests to elements of the storage cluster.

Examples include: HP (DEC) Tru64 cluster and Spinnaker is a clustered NAS (NFS) service. Panasas ActiveScale is a cluster filesystem

Parallel filesystem - file systems with support for parallel applications, all nodes may be accessing the same files at the same time, concurrent read and write. Examples of this include: Panasas ActiveScale, Lustre, GPFS and Sistina.

Finally, these definitions overlap. A SAN filesystem can be symmetric or asymmetric. Its servers can be clustered or single. And it can support parallel apps or not.

论坛徽章:
0
89 [报告]
发表于 2005-06-14 18:49 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

raidcracker 写到:
开发测试一个人做,要出乱子的.:>;


http://bbs.chinaunix.net/forum/viewtopic.php?t=544517&show_type=&postdays=0&postorder=asc&start=80

给我们的开发分工和流程提点建议如何?

------------------------------------------------------------

对项目管理我并没有很多的经验,只是深恶痛绝开发凌驾于测试之上.
我感觉你的分工就有这个倾向.测试应该是独立的并行于开发指导,而且测试是测试工程师的职责,调试是开发工程师的职责.不能和二为一.

最后不要误会,我不是搞测试的来为测试说话的,而是搞Raid和SAN的研发的.

论坛徽章:
0
90 [报告]
发表于 2005-06-16 00:13 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

按照甘特图的画法, 应该是这个样子的吧

  |
  |  研发 ->;
  |    ^ 开发 ->;
  |       ^ QA ->;
  |          ^ 运营(维护) ->;
----------------------------

那你对 FC  SCSI  iSCSI NFS SAMBA CACHE RAID 等很熟悉喽  
您需要登录后才可以回帖 登录 | 注册

本版积分规则 发表回复

  

北京盛拓优讯信息技术有限公司. 版权所有 京ICP备16024965号-6 北京市公安局海淀分局网监中心备案编号:11010802020122 niuxiaotong@pcpop.com 17352615567
未成年举报专区
中国互联网协会会员  联系我们:huangweiwei@itpub.net
感谢所有关心和支持过ChinaUnix的朋友们 转载本站内容请注明原作者名及出处

清除 Cookies - ChinaUnix - Archiver - WAP - TOP