是一个并行计算框架(计算的数据源比较广泛-HDFS、RDBMS、NoSQL),Hadoop的 MR模块充分利用了HDFS中所有数据节点(datanode)所在机器的内存、CUP以及少量磁盘完成对大数据集的分布式计算。MapReduce将计算分为两个阶段:

  1. 通过将一个大的计算任务分割成若干个小任务(计算目标数据集的分割),每一个小任务会分配给所有的计算节点(datanode所在物理机器)完成对局部数据的归类和分析,我们通常把该阶段定义为Map阶段,在Map阶段结束后会在本地系统磁盘存储计算的临时结果;
  2. 当Map阶段所有节点完成对局部数据的归类分析后,MR框架会启动Reduce任务完成对Map阶段的局部计算临时结果汇总,把以上阶段成为Reduce阶段。

I、计算流程

II、YARN环境搭建

配置文件

[root@CentOS ~]# vi /usr/hadoop-2.6.0/etc/hadoop/yarn-site.xml

<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<!--Resource Manager-->
<property>
<name>yarn.resourcemanager.hostname</name>
<value>CentOS</value>
</property>

[root@CentOS ~]# mv /usr/hadoop-2.6.0/etc/hadoop/mapred-site.xml.template /usr/hadoop-2.6.0/etc/hadoop/mapred-site.xml

[root@CentOS ~]# vi /usr/hadoop-2.6.0/etc/hadoop/mapred-site.xml

<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>

启动计算服务

[root@CentOS ~]# start-yarn.sh
[root@CentOS ~]# jps
1584 SecondaryNameNode
1364 NameNode
1446 DataNode
5229 Jps

访问:http://centos:8088/

III、HelloWorld of MapReduce 编程

  <dependency>
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-mapreduce-client-core</artifactId>
<version>2.6.0</version>
</dependency>
<dependency>
<groupId>org.apache.hadoop</groupId>
<artifactId>hadoop-mapreduce-client-jobclient</artifactId>
<version>2.6.0</version>
</dependency>

IpMapper

 import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Mapper; import java.io.IOException; /**
* @program: hadoop_01
* @description:
* @author: luoht
* @create: 2019-01-04 16:08
**/ public class IpMapper extends Mapper<LongWritable,Text,Text,IntWritable>{ /**
*192.168.0.12 1 001 click 5000 2019-01-04 14:44:00
* @param key :输入文本行字节偏移量
* @param value:输入文本行
* @param context
* @throws IOException
* @throws InterruptedException
*/ @Override
protected void map(LongWritable key, Text value, Context context) throws IOException, InterruptedException {
String[] tokens = value.toString().split("");
String ip = tokens[0];
context.write(new Text(ip),new IntWritable(1));
}
}

IpReducer


import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Reducer; import java.io.IOException; /**
* @program: hadoop_01
* @description:
* @author: luoht
* @create: 2019-01-04 16:13
**/ public class IpReducer extends Reducer<Text,IntWritable,Text,IntWritable> {
/**
*
* @param key :ip
* @param values: Int[]{1,1,1,..}
* @param context
* @throws IOException
* @throws InterruptedException
*/ @Override
protected void reduce(Text key, Iterable<IntWritable> values, Context context) throws IOException, InterruptedException {
int total = 0;
for (IntWritable value : values) {
total+=value.get();
}
context.write(key,new IntWritable(total)); }
}

封装job


import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.conf.Configured;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.lib.input.TextInputFormat;
import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
import org.apache.hadoop.util.Tool;
import org.apache.hadoop.util.ToolRunner; /**
* @program: hadoop_01
* @description:
* @author: luoht
* @create: 2019-01-04 16:15
**/ public class CustomJobSubmiter extends Configured implements Tool {
@Override
public int run(String[] strings) throws Exception {
/*1. 封装job 对象*/
Configuration conf=getConf();
Job job = Job.getInstance(conf);
/*2. 设置数据读入和写出的格式*/
job.setInputFormatClass(TextInputFormat.class);
job.setOutputFormatClass(TextOutputFormat.class);
/*3. 设置处理数据的路径*/
Path dst = new Path("/tt/test");
TextOutputFormat.setOutputPath(job,dst);
/*4. 设置数据计算逻辑*/
Path src=new Path("/tt/access");
TextInputFormat.addInputPath(job,src);
Path dst=new Path("/tt/result");
TextOutputFormat.setOutputPath(job,dst);
/*5. 设置Mapper和Reducer输出泛型*/
job.setMapOutputKeyClass(Text.class);
job.setMapOutputValueClass(IntWritable.class); job.setOutputKeyClass(Text.class);
job.setOutputValueClass(IntWritable.class);
/*6. 提交任务*/
job.submit();
return 0;
} public static void main(String[] args) throws Exception {
ToolRunner.run(new CustomJobSubmiter(),args);
}
}

IV、发布服务

1.远程 .jar包部署*

修改 job 服务

 public class CustomJobSubmiter extends Configured implements Tool {
@Override
public int run(String[] strings) throws Exception {
/*1. 封装job 对象*/
Configuration conf=getConf();
Job job = Job.getInstance(conf); /*设置jar类加载器*/
//设置jar类加载器,否则MapReduce框架找不到Mapper和Reducer
job.setJarByClass(CustomJobSubmiter.class); /*2. 设置数据读入和写出的格式*/
job.setInputFormatClass(TextInputFormat.class);
job.setOutputFormatClass(TextOutputFormat.class); /*3. 设置处理数据的路径*/
Path src=new Path("/tt/access");
TextInputFormat.addInputPath(job,src);
Path dst=new Path("/tt/result");
TextOutputFormat.setOutputPath(job,dst); /*4. 设置数据计算逻辑*/
job.setMapperClass(IpMapper.class);
job.setReducerClass(IpReducer.class); /*5. 设置Mapper和Reducer输出泛型*/
job.setMapOutputKeyClass(Text.class);
job.setMapOutputValueClass(IntWritable.class); job.setOutputKeyClass(Text.class);
job.setOutputValueClass(IntWritable.class); /*6. 提交任务 可查看过程*/ job.waitForCompletion(true);
return 0;
} public static void main(String[] args) throws Exception {
ToolRunner.run(new CustomJobSubmiter(),args);
}
}

上传jar包到linux 根部录下,将需要计算的文件上传到 /tt/test

[root@CentOS ~]# hdfs dfs -mkdir -p /tt/access
[root@CentOS ~]# vi t_access
[root@CentOS ~]# hdfs dfs -put /root/t_access /tt/access
[root@CentOS ~]# hadoop jar hadoop_mapreduce-1.0-SNAPSHOT.jar com.baizhi.mr.CustomJobSubmiter
[root@CentOS ~]# hdfs dfs -cat /tt/result/part-r-00000
1 4
2.本地模拟

添加日志文件


log4j.rootLogger=info,stdout
#console
log4j.appender.stdout=org.apache.log4j.ConsoleAppender
log4j.appender.stdout.layout=org.apache.log4j.PatternLayout
log4j.appender.stdout.layout.ConversionPattern=%p %d{yyyy-MM-dd HH:mm:ss a} %c %m%n

覆盖 org.apache.hadoop.io.nativeio.NativeIO

//
// Source code recreated from a .class file by IntelliJ IDEA
// (powered by Fernflower decompiler)
// package org.apache.hadoop.io.nativeio; import com.google.common.annotations.VisibleForTesting;
import java.io.Closeable;
import java.io.File;
import java.io.FileDescriptor;
import java.io.FileInputStream;
import java.io.FileOutputStream;
import java.io.IOException;
import java.io.RandomAccessFile;
import java.lang.reflect.Field;
import java.nio.ByteBuffer;
import java.nio.MappedByteBuffer;
import java.nio.channels.FileChannel;
import java.util.Map;
import java.util.concurrent.ConcurrentHashMap;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.classification.InterfaceAudience.Private;
import org.apache.hadoop.classification.InterfaceStability.Unstable;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.HardLink;
import org.apache.hadoop.io.IOUtils;
import org.apache.hadoop.io.SecureIOUtils.AlreadyExistsException;
import org.apache.hadoop.util.NativeCodeLoader;
import org.apache.hadoop.util.PerformanceAdvisory;
import org.apache.hadoop.util.Shell;
import sun.misc.Cleaner;
import sun.misc.Unsafe;
import sun.nio.ch.DirectBuffer; @Private
@Unstable
public class NativeIO {
private static boolean workaroundNonThreadSafePasswdCalls = false;
private static final Log LOG = LogFactory.getLog(NativeIO.class);
private static boolean nativeLoaded = false;
private static final Map<Long, CachedUid> uidCache;
private static long cacheTimeout;
private static boolean initialized; public NativeIO() {
} public static boolean isAvailable() {
return NativeCodeLoader.isNativeCodeLoaded() && nativeLoaded;
} private static native void initNative(); static long getMemlockLimit() {
return isAvailable() ? getMemlockLimit0() : 0L;
} private static native long getMemlockLimit0(); static long getOperatingSystemPageSize() {
try {
Field f = Unsafe.class.getDeclaredField("theUnsafe");
f.setAccessible(true);
Unsafe unsafe = (Unsafe)f.get((Object)null);
return (long)unsafe.pageSize();
} catch (Throwable var2) {
LOG.warn("Unable to get operating system page size. Guessing 4096.", var2);
return 4096L;
}
} private static String stripDomain(String name) {
int i = name.indexOf(92);
if (i != -1) {
name = name.substring(i + 1);
} return name;
} public static String getOwner(FileDescriptor fd) throws IOException {
ensureInitialized();
if (Shell.WINDOWS) {
String owner = Windows.getOwner(fd);
owner = stripDomain(owner);
return owner;
} else {
long uid = POSIX.getUIDforFDOwnerforOwner(fd);
CachedUid cUid = (CachedUid)uidCache.get(uid);
long now = System.currentTimeMillis();
if (cUid != null && cUid.timestamp + cacheTimeout > now) {
return cUid.username;
} else {
String user = POSIX.getUserName(uid);
LOG.info("Got UserName " + user + " for UID " + uid + " from the native implementation");
cUid = new CachedUid(user, now);
uidCache.put(uid, cUid);
return user;
}
}
} public static FileInputStream getShareDeleteFileInputStream(File f) throws IOException {
if (!Shell.WINDOWS) {
return new FileInputStream(f);
} else {
FileDescriptor fd = Windows.createFile(f.getAbsolutePath(), 2147483648L, 7L, 3L);
return new FileInputStream(fd);
}
} public static FileInputStream getShareDeleteFileInputStream(File f, long seekOffset) throws IOException {
if (!Shell.WINDOWS) {
RandomAccessFile rf = new RandomAccessFile(f, "r");
if (seekOffset > 0L) {
rf.seek(seekOffset);
} return new FileInputStream(rf.getFD());
} else {
FileDescriptor fd = Windows.createFile(f.getAbsolutePath(), 2147483648L, 7L, 3L);
if (seekOffset > 0L) {
Windows.setFilePointer(fd, seekOffset, 0L);
} return new FileInputStream(fd);
}
} public static FileOutputStream getCreateForWriteFileOutputStream(File f, int permissions) throws IOException {
FileDescriptor fd;
if (!Shell.WINDOWS) {
try {
fd = POSIX.open(f.getAbsolutePath(), 193, permissions);
return new FileOutputStream(fd);
} catch (NativeIOException var4) {
if (var4.getErrno() == Errno.EEXIST) {
throw new AlreadyExistsException(var4);
} else {
throw var4;
}
}
} else {
try {
fd = Windows.createFile(f.getCanonicalPath(), 1073741824L, 7L, 1L);
POSIX.chmod(f.getCanonicalPath(), permissions);
return new FileOutputStream(fd);
} catch (NativeIOException var3) {
if (var3.getErrorCode() == 80L) {
throw new AlreadyExistsException(var3);
} else {
throw var3;
}
}
}
} private static synchronized void ensureInitialized() {
if (!initialized) {
cacheTimeout = (new Configuration()).getLong("hadoop.security.uid.cache.secs", 14400L) * 1000L;
LOG.info("Initialized cache for UID to User mapping with a cache timeout of " + cacheTimeout / 1000L + " seconds.");
initialized = true;
} } public static void renameTo(File src, File dst) throws IOException {
if (!nativeLoaded) {
if (!src.renameTo(dst)) {
throw new IOException("renameTo(src=" + src + ", dst=" + dst + ") failed.");
}
} else {
renameTo0(src.getAbsolutePath(), dst.getAbsolutePath());
} } public static void link(File src, File dst) throws IOException {
if (!nativeLoaded) {
HardLink.createHardLink(src, dst);
} else {
link0(src.getAbsolutePath(), dst.getAbsolutePath());
} } private static native void renameTo0(String var0, String var1) throws NativeIOException; private static native void link0(String var0, String var1) throws NativeIOException; public static void copyFileUnbuffered(File src, File dst) throws IOException {
if (nativeLoaded && Shell.WINDOWS) {
copyFileUnbuffered0(src.getAbsolutePath(), dst.getAbsolutePath());
} else {
FileInputStream fis = null;
FileOutputStream fos = null;
FileChannel input = null;
FileChannel output = null; try {
fis = new FileInputStream(src);
fos = new FileOutputStream(dst);
input = fis.getChannel();
output = fos.getChannel();
long remaining = input.size();
long position = 0L; for(long transferred = 0L; remaining > 0L; position += transferred) {
transferred = input.transferTo(position, remaining, output);
remaining -= transferred;
}
} finally {
IOUtils.cleanup(LOG, new Closeable[]{output});
IOUtils.cleanup(LOG, new Closeable[]{fos});
IOUtils.cleanup(LOG, new Closeable[]{input});
IOUtils.cleanup(LOG, new Closeable[]{fis});
}
} } private static native void copyFileUnbuffered0(String var0, String var1) throws NativeIOException; static {
if (NativeCodeLoader.isNativeCodeLoaded()) {
try {
initNative();
nativeLoaded = true;
} catch (Throwable var1) {
PerformanceAdvisory.LOG.debug("Unable to initialize NativeIO libraries", var1);
}
} uidCache = new ConcurrentHashMap();
initialized = false;
} private static class CachedUid {
final long timestamp;
final String username; public CachedUid(String username, long timestamp) {
this.timestamp = timestamp;
this.username = username;
}
} public static class Windows {
public static final long GENERIC_READ = 2147483648L;
public static final long GENERIC_WRITE = 1073741824L;
public static final long FILE_SHARE_READ = 1L;
public static final long FILE_SHARE_WRITE = 2L;
public static final long FILE_SHARE_DELETE = 4L;
public static final long CREATE_NEW = 1L;
public static final long CREATE_ALWAYS = 2L;
public static final long OPEN_EXISTING = 3L;
public static final long OPEN_ALWAYS = 4L;
public static final long TRUNCATE_EXISTING = 5L;
public static final long FILE_BEGIN = 0L;
public static final long FILE_CURRENT = 1L;
public static final long FILE_END = 2L;
public static final long FILE_ATTRIBUTE_NORMAL = 128L; public Windows() {
} public static native FileDescriptor createFile(String var0, long var1, long var3, long var5) throws IOException; public static native long setFilePointer(FileDescriptor var0, long var1, long var3) throws IOException; private static native String getOwner(FileDescriptor var0) throws IOException; private static native boolean access0(String var0, int var1); public static boolean access(String path, AccessRight desiredAccess) throws IOException {
return true;
// return access0(path, desiredAccess.accessRight());
} public static native void extendWorkingSetSize(long var0) throws IOException; static {
if (NativeCodeLoader.isNativeCodeLoaded()) {
try {
NativeIO.initNative();
NativeIO.nativeLoaded = true;
} catch (Throwable var1) {
PerformanceAdvisory.LOG.debug("Unable to initialize NativeIO libraries", var1);
}
} } public static enum AccessRight {
ACCESS_READ(1),
ACCESS_WRITE(2),
ACCESS_EXECUTE(32); private final int accessRight; private AccessRight(int access) {
this.accessRight = access;
} public int accessRight() {
return this.accessRight;
}
}
} public static class POSIX {
public static final int O_RDONLY = 0;
public static final int O_WRONLY = 1;
public static final int O_RDWR = 2;
public static final int O_CREAT = 64;
public static final int O_EXCL = 128;
public static final int O_NOCTTY = 256;
public static final int O_TRUNC = 512;
public static final int O_APPEND = 1024;
public static final int O_NONBLOCK = 2048;
public static final int O_SYNC = 4096;
public static final int O_ASYNC = 8192;
public static final int O_FSYNC = 4096;
public static final int O_NDELAY = 2048;
public static final int POSIX_FADV_NORMAL = 0;
public static final int POSIX_FADV_RANDOM = 1;
public static final int POSIX_FADV_SEQUENTIAL = 2;
public static final int POSIX_FADV_WILLNEED = 3;
public static final int POSIX_FADV_DONTNEED = 4;
public static final int POSIX_FADV_NOREUSE = 5;
public static final int SYNC_FILE_RANGE_WAIT_BEFORE = 1;
public static final int SYNC_FILE_RANGE_WRITE = 2;
public static final int SYNC_FILE_RANGE_WAIT_AFTER = 4;
private static final Log LOG = LogFactory.getLog(NativeIO.class);
private static boolean nativeLoaded = false;
private static boolean fadvisePossible = true;
private static boolean syncFileRangePossible = true;
static final String WORKAROUND_NON_THREADSAFE_CALLS_KEY = "hadoop.workaround.non.threadsafe.getpwuid";
static final boolean WORKAROUND_NON_THREADSAFE_CALLS_DEFAULT = true;
private static long cacheTimeout = -1L;
private static CacheManipulator cacheManipulator = new CacheManipulator();
private static final Map<Integer, CachedName> USER_ID_NAME_CACHE;
private static final Map<Integer, CachedName> GROUP_ID_NAME_CACHE;
public static final int MMAP_PROT_READ = 1;
public static final int MMAP_PROT_WRITE = 2;
public static final int MMAP_PROT_EXEC = 4; public POSIX() {
} public static CacheManipulator getCacheManipulator() {
return cacheManipulator;
} public static void setCacheManipulator(CacheManipulator cacheManipulator) {
cacheManipulator = cacheManipulator;
} public static boolean isAvailable() {
return NativeCodeLoader.isNativeCodeLoaded() && nativeLoaded;
} private static void assertCodeLoaded() throws IOException {
if (!isAvailable()) {
throw new IOException("NativeIO was not loaded");
}
} public static native FileDescriptor open(String var0, int var1, int var2) throws IOException; private static native Stat fstat(FileDescriptor var0) throws IOException; private static native void chmodImpl(String var0, int var1) throws IOException; public static void chmod(String path, int mode) throws IOException {
if (!Shell.WINDOWS) {
chmodImpl(path, mode);
} else {
try {
chmodImpl(path, mode);
} catch (NativeIOException var3) {
if (var3.getErrorCode() == 3L) {
throw new NativeIOException("No such file or directory", Errno.ENOENT);
} LOG.warn(String.format("NativeIO.chmod error (%d): %s", var3.getErrorCode(), var3.getMessage()));
throw new NativeIOException("Unknown error", Errno.UNKNOWN);
}
} } static native void posix_fadvise(FileDescriptor var0, long var1, long var3, int var5) throws NativeIOException; static native void sync_file_range(FileDescriptor var0, long var1, long var3, int var5) throws NativeIOException; static void posixFadviseIfPossible(String identifier, FileDescriptor fd, long offset, long len, int flags) throws NativeIOException {
if (nativeLoaded && fadvisePossible) {
try {
posix_fadvise(fd, offset, len, flags);
} catch (UnsupportedOperationException var8) {
fadvisePossible = false;
} catch (UnsatisfiedLinkError var9) {
fadvisePossible = false;
}
} } public static void syncFileRangeIfPossible(FileDescriptor fd, long offset, long nbytes, int flags) throws NativeIOException {
if (nativeLoaded && syncFileRangePossible) {
try {
sync_file_range(fd, offset, nbytes, flags);
} catch (UnsupportedOperationException var7) {
syncFileRangePossible = false;
} catch (UnsatisfiedLinkError var8) {
syncFileRangePossible = false;
}
} } static native void mlock_native(ByteBuffer var0, long var1) throws NativeIOException; static void mlock(ByteBuffer buffer, long len) throws IOException {
assertCodeLoaded();
if (!buffer.isDirect()) {
throw new IOException("Cannot mlock a non-direct ByteBuffer");
} else {
mlock_native(buffer, len);
}
} public static void munmap(MappedByteBuffer buffer) {
if (buffer instanceof DirectBuffer) {
Cleaner cleaner = ((DirectBuffer)buffer).cleaner();
cleaner.clean();
} } private static native long getUIDforFDOwnerforOwner(FileDescriptor var0) throws IOException; private static native String getUserName(long var0) throws IOException; public static Stat getFstat(FileDescriptor fd) throws IOException {
Stat stat = null;
if (!Shell.WINDOWS) {
stat = fstat(fd);
stat.owner = getName(IdCache.USER, stat.ownerId);
stat.group = getName(IdCache.GROUP, stat.groupId);
} else {
try {
stat = fstat(fd);
} catch (NativeIOException var3) {
if (var3.getErrorCode() == 6L) {
throw new NativeIOException("The handle is invalid.", Errno.EBADF);
} LOG.warn(String.format("NativeIO.getFstat error (%d): %s", var3.getErrorCode(), var3.getMessage()));
throw new NativeIOException("Unknown error", Errno.UNKNOWN);
}
} return stat;
} private static String getName(IdCache domain, int id) throws IOException {
Map<Integer, CachedName> idNameCache = domain == IdCache.USER ? USER_ID_NAME_CACHE : GROUP_ID_NAME_CACHE;
CachedName cachedName = (CachedName)idNameCache.get(id);
long now = System.currentTimeMillis();
String name;
if (cachedName != null && cachedName.timestamp + cacheTimeout > now) {
name = cachedName.name;
} else {
name = domain == IdCache.USER ? getUserName(id) : getGroupName(id);
if (LOG.isDebugEnabled()) {
String type = domain == IdCache.USER ? "UserName" : "GroupName";
LOG.debug("Got " + type + " " + name + " for ID " + id + " from the native implementation");
} cachedName = new CachedName(name, now);
idNameCache.put(id, cachedName);
} return name;
} static native String getUserName(int var0) throws IOException; static native String getGroupName(int var0) throws IOException; public static native long mmap(FileDescriptor var0, int var1, boolean var2, long var3) throws IOException; public static native void munmap(long var0, long var2) throws IOException; static {
if (NativeCodeLoader.isNativeCodeLoaded()) {
try {
Configuration conf = new Configuration();
NativeIO.workaroundNonThreadSafePasswdCalls = conf.getBoolean("hadoop.workaround.non.threadsafe.getpwuid", true);
NativeIO.initNative();
nativeLoaded = true;
cacheTimeout = conf.getLong("hadoop.security.uid.cache.secs", 14400L) * 1000L;
LOG.debug("Initialized cache for IDs to User/Group mapping with a cache timeout of " + cacheTimeout / 1000L + " seconds.");
} catch (Throwable var1) {
PerformanceAdvisory.LOG.debug("Unable to initialize NativeIO libraries", var1);
}
} USER_ID_NAME_CACHE = new ConcurrentHashMap();
GROUP_ID_NAME_CACHE = new ConcurrentHashMap();
} private static enum IdCache {
USER,
GROUP; private IdCache() {
}
} private static class CachedName {
final long timestamp;
final String name; public CachedName(String name, long timestamp) {
this.name = name;
this.timestamp = timestamp;
}
} public static class Stat {
private int ownerId;
private int groupId;
private String owner;
private String group;
private int mode;
public static final int S_IFMT = 61440;
public static final int S_IFIFO = 4096;
public static final int S_IFCHR = 8192;
public static final int S_IFDIR = 16384;
public static final int S_IFBLK = 24576;
public static final int S_IFREG = 32768;
public static final int S_IFLNK = 40960;
public static final int S_IFSOCK = 49152;
public static final int S_IFWHT = 57344;
public static final int S_ISUID = 2048;
public static final int S_ISGID = 1024;
public static final int S_ISVTX = 512;
public static final int S_IRUSR = 256;
public static final int S_IWUSR = 128;
public static final int S_IXUSR = 64; Stat(int ownerId, int groupId, int mode) {
this.ownerId = ownerId;
this.groupId = groupId;
this.mode = mode;
} Stat(String owner, String group, int mode) {
if (!Shell.WINDOWS) {
this.owner = owner;
} else {
this.owner = NativeIO.stripDomain(owner);
} if (!Shell.WINDOWS) {
this.group = group;
} else {
this.group = NativeIO.stripDomain(group);
} this.mode = mode;
} public String toString() {
return "Stat(owner='" + this.owner + "', group='" + this.group + "'" + ", mode=" + this.mode + ")";
} public String getOwner() {
return this.owner;
} public String getGroup() {
return this.group;
} public int getMode() {
return this.mode;
}
} @VisibleForTesting
public static class NoMlockCacheManipulator extends CacheManipulator {
public NoMlockCacheManipulator() {
} public void mlock(String identifier, ByteBuffer buffer, long len) throws IOException {
POSIX.LOG.info("mlocking " + identifier);
} public long getMemlockLimit() {
return 1125899906842624L;
} public long getOperatingSystemPageSize() {
return 4096L;
} public boolean verifyCanMlock() {
return true;
}
} @VisibleForTesting
public static class CacheManipulator {
public CacheManipulator() {
} public void mlock(String identifier, ByteBuffer buffer, long len) throws IOException {
POSIX.mlock(buffer, len);
} public long getMemlockLimit() {
return NativeIO.getMemlockLimit();
} public long getOperatingSystemPageSize() {
return NativeIO.getOperatingSystemPageSize();
} public void posixFadviseIfPossible(String identifier, FileDescriptor fd, long offset, long len, int flags) throws NativeIOException {
POSIX.posixFadviseIfPossible(identifier, fd, offset, len, flags);
} public boolean verifyCanMlock() {
return NativeIO.isAvailable();
}
}
}
}
3.跨平台测试

更新 job

 public class CustomJobSubmiter extends Configured implements Tool {
@Override
public int run(String[] args) throws Exception { //1.封装job对象
Configuration conf=getConf();
conf.addResource("core-site.xml");
conf.addResource("hdfs-site.xml");
conf.addResource("yarn-site.xml");
conf.addResource("mapred-site.xml");
conf.set(MRJobConfig.JAR,"file:///E:\\BD139\\DBcode\\hadoop_crossPlatFrom\\target\\hadoop_crossPlatFrom-1.0-SNAPSHOT.jar");
Job job=Job.getInstance(conf); //2.设置数据读入和写出格式
job.setInputFormatClass(TextInputFormat.class);
job.setOutputFormatClass(TextOutputFormat.class); //3.设置处理数据路径
Path src=new Path("/tt/access");
TextInputFormat.addInputPath(job,src);
Path dst=new Path("/tt/result");
TextOutputFormat.setOutputPath(job,dst); //4.设置数据计算逻辑
job.setMapperClass(IpMapper.class);
job.setReducerClass(IpReducer.class); //5.设置Mapper和Reducer输出泛型
job.setMapOutputKeyClass(Text.class);
job.setMapOutputValueClass(IntWritable.class); job.setOutputKeyClass(Text.class);
job.setOutputValueClass(IntWritable.class); //6.提交任务
//job.submit();
job.waitForCompletion(true);
return 0;
} public static void main(String[] args) throws Exception {
ToolRunner.run(new CustomJobSubmiter(),args);
}
}

引入 Hadoop 配置文件

mapred-site.xml 加入 支持跨平台 配置

<property>
<name>mapreduce.app-submission.cross-platform</name>
<value>true</value>
</property>

五、MapReduce 发布服务的更多相关文章

  1. ArcGIS API for javascript开发笔记(五)——GP服务调用之GP模型的发布及使用详解

    感谢一路走来默默陪伴和支持的你~~~ ----------------欢迎来访,拒绝转载---------------- 关于GP模型的制作请点我! 一.GP发布 ArcGIS Desktop可以作为 ...

  2. ServiceStack.Redis订阅发布服务的调用(Z)

      1.Redis订阅发布介绍Redis订阅发布是一种消息通信模式:发布者(publisher)发送消息,订阅者(Subscriber)接受消息.类似于设计模式中的观察者模式.发布者和订阅者之间使用频 ...

  3. ArcGIS Server开发教程系列(2)配置ARCMAP和ARCCatalog发布服务

    1.       Arc catalog的配置 打开catalog,如图新增刚刚创建的server 1. Use GIS services: 用户身份连接 使用此种连接,可以浏览.使用站点内发布的所有 ...

  4. ServiceStack.Redis订阅发布服务的调用

    1.Redis订阅发布介绍 Redis订阅发布是一种消息通信模式:发布者(publisher)发送消息,订阅者(Subscriber)接受消息.类似于设计模式中的观察者模式. 发布者和订阅者之间使用频 ...

  5. ArcGIS发布服务时缓存切片设置

    [文件]>[共享]>[服务]>[覆盖原有服务]或[创建新服务] 设置好相关参数后,会弹出"服务编辑框": 进入"缓存" 1."绘制此 ...

  6. 使用expect脚本语言写一键发布服务(代码发布、所有服务重启)

    互联网服务有很多台服务,但是在上线的时候需要将这些服务版本都更新与个个都重启,下面的脚本语言,就是一键发布服务~ 1.在/home/weihu/deploy/ 目录下建下publish .publis ...

  7. PHPRPC jsp发布服务

    PHPRPC是面向服务的高性能远程过程调用协议.PHPRPC for java 是此协议的一种实现,详细请到官方看PHPRPC的介绍 ,以其它rpc协议的性能对比请:Java.PHPRPC.Hessi ...

  8. 除非 Windows Activation Service (WAS)和万维网发布服务(W3SVC)均处于运行状态,否则无法启动网站。目前,这两项服务均处于停止状态。

    win7 IIS 所有网站都停止了,启动提示: 除非 Windows Activation Service (WAS)和万维网发布服务(W3SVC)均处于运行状态,否则无法启动网站.目前,这两项服务均 ...

  9. arcgis server10.1注册服务——避免在发布服务中拷贝数据

    之前用的arcgis10.1前的版本,后来换成10.1还有点不习惯,变化挺多的.发布服务过程中,进行分析的时候,其中有一项提醒:xxx图层没有注册到服务,很纳闷,为什么会有这种提示,不管,点击发布,会 ...

随机推荐

  1. 从Pc转向H5开发遇到的适配问题思考

    1.首先说滚动条 移动端开发在不设置任何适配和viewport宽度的情况下,以iphone5为例:屏幕界面的逻辑分辨率是320x568,在谷歌浏览器的界面下屏幕的可视宽度是980px(谷歌设置的,每个 ...

  2. HTML学习笔记《二》 ----HTML部分标签应用

    HTML 部分标签应用 一.HTML中的注释.标签语法以及颜色表示方式 1.HTML中注释: <!-- .... --> 注释的目的:1. 辅助说明 2.代码调试(不同编程语言,注释方式略 ...

  3. php基础部分(1)

    PHP 输出文本的基础指令:echo 和 print.echo和print的区别echo是PHP语句, print和print_r是函数,语句没有返回值,函数可以有返回值(即便没有用)echo 输出一 ...

  4. arcgis-Pro-1.3

    sdk在线地址 http://pro.arcgis.com/en/pro-app/sdk/api-reference/#topic14079.html 看了下,整体开发架构全变了. 觉得不如以前com ...

  5. 139.00.005 Git学习-分支管理

    @(139 - Environment Settings | 环境配置) 一.Why? 分支在实际中有什么用呢?假设你准备开发一个新功能,但是需要两周才能完成,第一周你写了50%的代码,如果立刻提交, ...

  6. MongoDB安装步骤

    安装C:\Users\Administrator>d:\mongo\bin\mongod -dbpath=D:\ND.Monodb\ND.Monodb.db -logpath=D:\ND.Mon ...

  7. 关于微信公账号H5 API 调用的坑 BUG

    页面A已经配置过,如果是单页面跳转,则页面B可以共享当前的SDK配置(至少菜单是这样的) 刷新页面,原先的菜单仍然会保持原样,只是调用SDK已经失效了,需要重新配置,重新配置后,菜单仍然会保持原样(如 ...

  8. C++的extern关键字

    extern是一个声明,不是一个定义,A模块想应用B模块的一个函数或者变量,A模块包含B模块的头文件,并且在变量或者头文件前,加 extern,虽然编译的时候,找不到模块的定义,但是在连接的时候,会在 ...

  9. 如何退出virtualbox scale mode

    进入scale mode之后,可能会退不出来 HOST Key + C. 默认是 右Ctrl + C

  10. vim c++插件clang_complete

    地址 http://www.vim.org/scripts/script.php?script_id=3302 以后应该会用到的,先记下:需要安装clang编译器.