I am working on a search engine design, which is to be run on cloud. We have just started, and have not much idea about Hdoop. Can anyone tell if HBase , MapReduce and HDFS can work on a single machine having Hdoop installed and running on it ?
相关问题
- What is the best way to do a search in a large fil
- Search Multiple Arrays for
- What are the advantages ManageIQ has over OpenStac
- Unable to generate jar file for Hadoop
- Find index given multiple values of array with Num
相关文章
- What is the complexity of bisect algorithm?
- mapreduce count example
- Visual Studio: Is there an incremental search for
- hbase-client 2.0.x error
- How do I hide a site from search engines? [closed]
- Why is C# Array.BinarySearch so fast?
- Find three elements in a sorted array which sum to
- Exception in thread “main” java.lang.NoClassDefFou
same here, I am running hadoop/hbase/hive on a single computer. If you really really want to see distributed computing on a single computer, grab lots of RAM, some hard disk space and go like this -
this could just be an experiment, because unless you have a decent multi-cpu or multi-core system, such a configuration will actually consume more on maintaining itself than giving you any performance.
gud luck.
--l4l
Yes you can. You can even create a Virtual Machine and run it on there on a single "computer" (which is what I have :) ).
The key is to simply install Hadoop in "Pseudo Distributed Mode" which is even described in the Hadoop Quickstart.
If you use the Cloudera distribution they have even created the configs needed for that in an RPM. Look here for more info in that.
HTH
Yes. In my development environment, I run
In addition, I run my applications, and map and reduce tasks launched by the task tracker. Running so many processes on the same machine results in a lot of contention for CPU cores, memory, and disk I/O, so it's definitely not great for high performance, but there is no limitation other than the amount of resources available.