免费注册 查看新帖 |

Chinaunix

  平台 论坛 博客 文库
最近访问板块 发新帖
楼主: yftty
打印 上一主题 下一主题

浩存 - 面向数据库,虚拟机等海量数据可同时提供NFS/iSCSI访问的集群存储系统 [复制链接]

论坛徽章:
0
161 [报告]
发表于 2005-08-26 22:26 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

[quote]原帖由 "suran007"]先谢谢楼主了,我看完之后,还是一头雾水,他讲的都是一些概念上的东东,要是有一些作多gnbd server的例子就好了,但是怎么也找不到,毕竟做这块的人还是很少的,郁闷中~~~~~[/quote 发表:


理论上可行,但会导致可用性的下降,如下面邮件所述:

https://www.redhat.com/archives/linux-cluster/2005-August/msg00270.html

论坛徽章:
0
162 [报告]
发表于 2005-08-29 16:41 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

我是计算机系的在读研究生,做的就是cluster fs这方面的工作
搞过一段时间的pvfs2,现在重点在研究lustre

在网上查资料的时候,看到了这个项目,
比较感兴趣,就把所有的回复都翻了一下,我想我也可以做点工作

不知道这个项目现在进展到什么程度,我可以参予吗?

论坛徽章:
0
163 [报告]
发表于 2005-08-30 13:02 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

原帖由 "artifly" 发表:
我是计算机系的在读研究生,做的就是cluster fs这方面的工作
搞过一段时间的pvfs2,现在重点在研究lustre

在网上查资料的时候,看到了这个项目,
比较感兴趣,就把所有的回复都翻了一下,我想我也可以做点工作
..........

"把所有的回复都翻了一下", 很好 -->; 我LP问我贴的这些东西我都看过没有, 呵呵,这些我都是一个字一个字看过来才贴上来的
第一版本这个月就可以发布了.
找个时间来公司一起聊聊吧先.

论坛徽章:
0
164 [报告]
发表于 2005-08-31 10:14 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

in Coda Lists

On Mon, Aug 15, 2005 at 01:31:43PM -0400, Kris Maglione wrote:
>; There's nothing stopping Coda (in theory. I haven't seen the code
>; relating to this) from implementing both partial and full file caching.
>; Whether it be a knob between two modes of caching, a switch to require
>; the fetching of all blocks (with unneeded ones at a lower priority, put
>; off until essential data is retrieved), or just a program like hoard
>; deciding what files need to be cached fully, and doing so. I'm not
>; saying that this should or will be implemented, but it is possible, in
>; theory. For Coda and AFS.

Actually there are many reasons to not have block level caching in Coda.

- VM deadlocks
   Because we have a userspace cache manager we could get into the
   situation where we are told to write out dirty data, but this causes
   us to request one or more memory pages from the kernel, either
   because we allocate memory, or are simply paging in some of the
   application/library code.  The kernel might then decide to give us
   pages that would require write-back of more dirty state to the
   userspace daemon. We would have to push venus into the kernel, which
   is what AFS did, but they aren't dealing with a lot of the same
   complexities like replication and reintegration.

- Code complexity
   It is already a hard enough problem to do optimistic replication and
   reintegration with whole files. The last thing I need right now is
   to add additional complexity so we suddenly have to reason about
   situations where we only happen to have parts of a locally modified
   file, which might already have been partially reintegration, but
   then overwritten on the server by another client and how to commit,
   revert or merge these local changes in the global replica(s). As
   well as effectivly maintaining the required data structures. The
   current RVM limitations are on number of file objects and not
   dependent on file size. You can cache 100 zero length files with the
   same overhead as far as the client in concerned as 100 files that
   are 1GB in size.

- Network performance
   It is more efficient to fetch a large file at once compared to
   requesting individual blocks. Available network bandwidth keeps
   increasing, but latency is bounded by the laws of physics. So the
   60ms roundtrip from coast-to-coast will remain. So requesting 1000
   individual 4KB blocks will always cost at least 60 seconds, while
   fetching a the same 4MB as a single file will become cheaper over
   time.

- Local performance
   Handling upcalls is quite expensive, there are at least 2 context
   switches and possibly some swapping/paging involved to get the
   request up to the cache manager and the response back to the
   application. Doing this on individual read and write operations
   would make the system a lot less responsive.

- Consistency model
   It is really easy to explain Coda's consistency model wrt other
   clients. You fetch a copy of the file when it is opened, and it is
   written back to the servers when it is closed (and it was modified).
   Now try to do the same if the client uses block-level caching. The
   picture quickly becomes very blurry, and Transarc AFS actually had
   (has?) a serious bug that leads to unexpected data loss in this area
   if people were assuming that it actually still provides AFS semantics.

   Also once a system provides block-level access, people start to
   expect the file system provides something close to UNIX semantics,
   which is really not a very usable model for any distributed
   filesystem.

Jan

论坛徽章:
0
165 [报告]
发表于 2005-09-02 14:54 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

我正在做lustre的试验,用的是1.0.4的版本,
我的内核是2.4.20-28.9_lustre.1.0.4smp
我的拓扑是一个OST(n01)(OSS),一个MDS(n03),一个CLIENT(n04)
我想在已有OST的OSS上再加入一个ost(n02)(oss),还要保留原来在oss上保存的数据,
当我在MDS上执行lconf --write_conf newconfig.xml让MDS从新读取新的oss的信息,然后继续启动MDS,却出现如下错误:
MDSDEV: n03-mds1 n03-mds1_UUID /dev/sda1 ext3 200000 no
Lustre:2195 (fsfilt_ext3.c:815:fsfilt_ext3_setup () ) Enable PDIROPS
LustreError:2195obd_config.c:101:class_attach ()) obd OSC_n03_n01-ost1_n03-mds 1 already attached
LustreError:2197obd_config.c:285:class_cleanup ()) Device 1 not setup ! /usr/sbin/lctl (17):error:setup: File exists
这是为什么?我要是用lconf --reformat --node n03 newconfig.xml就可以启动mds,client也可挂载,挂在分区的容量已经变为两个ost容量的合计了,但是这样就无法保存我原来在/mnt/lustre/中的数据了
那位朋友做过lustre的,请帮帮我,我的qq24242546
以下是我的newconfig.sh脚本内容:
#!/bin/sh

#config.sh

#Create nodes
rm -f newconfig.xml
lmc -m newconfig.xml --add net --node n03 --nid n03 --nettype tcp
lmc -m newconfig.xml --add net --node n01 --nid n01 --nettype tcp
lmc -m newconfig.xml --add net --node n02 --nid n02 --nettype tcp
lmc -m newconfig.xml --add net --node client --nid '*' --nettype tcp

#Configure mds
lmc -m newconfig.xml --add mds --node n03 --mds n03-mds1 --fstype ext3 --dev /dev/sda1 --size 200000

#Configure ost
lmc -m newconfig.xml --add lov --lov lov1 --mds n03-mds1 --stripe_sz 1048576 --stripe_cnt 0 --stripe_pattern 0
lmc -m newconfig.xml --add ost --node n01 --lov lov1 --ost n01-ost1 --fstype ext3 --dev /dev/sda1 --size 1000000
lmc -m newconfig.xml --add ost --node n02 --lov lov1 --ost n02-ost1 --fstype ext3 --dev /dev/sda1 --size 1000000

#Configure client
lmc -m newconfig.xml --add mtpt --node client --path /mnt/lustre --mds n03-mds1 --lov lov1

论坛徽章:
0
166 [报告]
发表于 2005-09-06 11:22 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

小弟前段时间翻译了GoogleFS的白皮书,不太完全,看到许多弟兄都有兴趣,贴出来与大家一起翻吧。译文比较多,另开了一个页面:
http://bbs.chinaunix.net/forum/viewtopic.php?show_type=&p=4054433#4054433

另向游兄问好,感谢关于FS体系结构方面的指导。

论坛徽章:
0
167 [报告]
发表于 2005-09-06 11:56 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

hehe, you did a good job
I'll read it line to line

论坛徽章:
0
168 [报告]
发表于 2005-09-06 21:11 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

Thanks a lot ,
these days I'm preparing my graduation thesis, I hate write doc so it's a terrible work for me, God save me!

论坛徽章:
0
169 [报告]
发表于 2005-09-12 18:39 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

初步看了googlefs的文档,还是有一些疑问:
1. 怎么使得文件大小为64M.
2.  master 怎么解决主备?
3. 如果解决chunk server的负载均衡问题?

请yftty兄回答,谢谢。

论坛徽章:
0
170 [报告]
发表于 2005-09-12 22:24 |只看该作者

Unix下针对邮件,搜索,网络硬盘等海量存储的分布式文件系统项目

原帖由 "szfrank" 发表:
初步看了googlefs的文档,还是有一些疑问:
1. 怎么使得文件大小为64M.
2.  master 怎么解决主备?
3. 如果解决chunk server的负载均衡问题?

请yftty兄回答,谢谢。


1. 怎么使得文件大小为64M.

It's not that the file size is 64MB, but the chunk size is 64MB. So as many little files can be spliced into one chunk, and big file can be splitted into multiple chunks.

2.  master 怎么解决主备?

"If its machine or disk fails, monitoring infrastructure outside GFS starts a new master process elsewhere with the replicated operation log."

3. 如果解决chunk server的负载均衡问题?

"4.3 Creation, Re-replication, Rebalancing

Chunk replicas are created for three reasons: chunk creation, re-replication, and rebalancing.
...
Finally, the master rebalances replicas periodically: it examines the current replica distribution and moves replicas for better disk space and load balancing. ..."
您需要登录后才可以回帖 登录 | 注册

本版积分规则 发表回复

  

北京盛拓优讯信息技术有限公司. 版权所有 京ICP备16024965号-6 北京市公安局海淀分局网监中心备案编号:11010802020122 niuxiaotong@pcpop.com 17352615567
未成年举报专区
中国互联网协会会员  联系我们:huangweiwei@itpub.net
感谢所有关心和支持过ChinaUnix的朋友们 转载本站内容请注明原作者名及出处

清除 Cookies - ChinaUnix - Archiver - WAP - TOP