-
Notifications
You must be signed in to change notification settings - Fork 34
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Create AbusiveUserMapper to replicate an abusive user's workload #94
Open
csgregorian
wants to merge
3
commits into
linkedin:master
Choose a base branch
from
csgregorian:cgregori-big-file-mapper
base: master
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
Open
Changes from all commits
Commits
File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
172 changes: 172 additions & 0 deletions
172
...-workload/src/main/java/com/linkedin/dynamometer/workloadgenerator/AbusiveUserMapper.java
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,172 @@ | ||
/** | ||
* Copyright 2019 LinkedIn Corporation. All rights reserved. Licensed under the BSD-2 Clause license. | ||
* See LICENSE in the project root for license information. | ||
*/ | ||
package com.linkedin.dynamometer.workloadgenerator; | ||
|
||
import org.apache.hadoop.conf.Configuration; | ||
import org.apache.hadoop.fs.FileSystem; | ||
import org.apache.hadoop.fs.Path; | ||
import org.apache.hadoop.hdfs.DistributedFileSystem; | ||
import org.apache.hadoop.io.LongWritable; | ||
import org.apache.hadoop.io.NullWritable; | ||
import org.apache.hadoop.security.UserGroupInformation; | ||
|
||
import java.io.IOException; | ||
import java.net.URI; | ||
import java.security.PrivilegedAction; | ||
import java.util.List; | ||
import java.util.concurrent.ConcurrentHashMap; | ||
|
||
|
||
/** | ||
* AbusiveUserMapper consists of two types of users: one type that performs a large listing operation on a directory | ||
* with many files, and another type that performs a small operation (either a single file listing or a single directory | ||
* creation). This uses {@link TimedInputFormat}; see its Javadoc for configuration information. Only requires the | ||
* NameNode, does not make file data changes. | ||
* | ||
* <p>Configuration options available: | ||
* <ul> | ||
* <li>{@value FILE_PARENT_PATH_KEY} (default: {@value FILE_PARENT_PATH_DEFAULT}): The root directory for the job to | ||
* create files in.</li> | ||
* <li>{@value NUM_LARGE_JOBS_KEY} (default: {@value NUM_LARGE_JOBS_DEFAULT}): The number of large listing jobs. | ||
* Mappers numbered 0 through this value will run the abusive operation.</li> | ||
* <li>{@value NUM_FILES_KEY} (default: {@value NUM_FILES_DEFAULT}): Number of files in the directory for the large | ||
* listing operations.</li> | ||
* <li>{@value MAPPERS_PER_USER_KEY} (default: {@value MAPPERS_PER_USER_DEFAULT}): Number of mappers to assign to each | ||
* user. Make a single large/abusive user by setting this and {@value NUM_LARGE_JOBS_KEY} equal. | ||
* <li>{@value ENABLE_WRITE_OPS_KEY} (default: {@value ENABLE_WRITE_OPS_DEFAULT}): Enabling this sets the small | ||
* operation to create a single directory, instead of listing a single file directory. | ||
* </ul> | ||
*/ | ||
public class AbusiveUserMapper extends WorkloadMapper<LongWritable, NullWritable, NullWritable, NullWritable> { | ||
|
||
public static final String FILE_PARENT_PATH_KEY = "bigfile.file-parent-path"; | ||
public static final String FILE_PARENT_PATH_DEFAULT = "/tmp/createFileMapper"; | ||
|
||
public static final String NUM_LARGE_JOBS_KEY = "bigfile.num-large-jobs"; | ||
public static final int NUM_LARGE_JOBS_DEFAULT = 10; | ||
|
||
public static final String NUM_FILES_KEY = "bigfile.num-files"; | ||
public static final int NUM_FILES_DEFAULT = 1000; | ||
|
||
public static final String MAPPERS_PER_USER_KEY = "bigfile.mappers-per-user"; | ||
public static final int MAPPERS_PER_USER_DEFAULT = 10; | ||
|
||
public static final String ENABLE_WRITE_OPS_KEY = "bigfile.enable-write-ops"; | ||
public static final boolean ENABLE_WRITE_OPS_DEFAULT = false; | ||
|
||
private ConcurrentHashMap<String, FileSystem> fsMap = new ConcurrentHashMap<String, FileSystem>(); | ||
private Path parentFolder; | ||
private int numLargeJobs; | ||
private int numFiles; | ||
private int mappersPerUser; | ||
private boolean enableWriteOps; | ||
|
||
@Override | ||
public String getDescription() { | ||
return "This mapper creates a number of large and small operations belonging to different users."; | ||
} | ||
|
||
@Override | ||
public List<String> getConfigDescriptions() { | ||
List<String> baseList = TimedInputFormat.getConfigDescriptions(); | ||
baseList.add(FILE_PARENT_PATH_KEY + " (default: " + FILE_PARENT_PATH_DEFAULT + | ||
"): The root directory for the job to create files in."); | ||
baseList.add(NUM_LARGE_JOBS_KEY + " (default: " + NUM_LARGE_JOBS_DEFAULT + | ||
"): Number of large listing jobs."); | ||
baseList.add(NUM_FILES_KEY + " (default: " + NUM_FILES_DEFAULT + | ||
"): Number of files in the directory for the large listing operations."); | ||
baseList.add(MAPPERS_PER_USER_KEY + " (default: " + MAPPERS_PER_USER_DEFAULT + | ||
"): Number of mappers per user."); | ||
baseList.add(ENABLE_WRITE_OPS_KEY + " (default: " + ENABLE_WRITE_OPS_DEFAULT + | ||
"): For the small operation, creates a single directory instead of listing a single file directory."); | ||
return baseList; | ||
} | ||
|
||
@Override | ||
public boolean verifyConfigurations(Configuration conf) { | ||
return TimedInputFormat.verifyConfigurations(conf); | ||
} | ||
|
||
@Override | ||
public void setup(Context context) throws IOException { | ||
// Load config | ||
Configuration conf = context.getConfiguration(); | ||
int taskID = context.getTaskAttemptID().getTaskID().getId(); | ||
numLargeJobs = conf.getInt(NUM_LARGE_JOBS_KEY, NUM_LARGE_JOBS_DEFAULT); | ||
numFiles = conf.getInt(NUM_FILES_KEY, NUM_FILES_DEFAULT); | ||
mappersPerUser = conf.getInt(MAPPERS_PER_USER_KEY, MAPPERS_PER_USER_DEFAULT); | ||
enableWriteOps = conf.getBoolean(ENABLE_WRITE_OPS_KEY, ENABLE_WRITE_OPS_DEFAULT); | ||
|
||
// Load filesystem | ||
String namenodeURI = conf.get(WorkloadDriver.NN_URI); | ||
UserGroupInformation loginUser = UserGroupInformation.getLoginUser(); | ||
String proxyUser = "fakeuser" + taskID / mappersPerUser; | ||
UserGroupInformation ugi = UserGroupInformation.createProxyUser(proxyUser, loginUser); | ||
FileSystem fs = ugi.doAs(new PrivilegedAction<FileSystem>() { | ||
@Override | ||
public FileSystem run() { | ||
try { | ||
FileSystem fs = new DistributedFileSystem(); | ||
fs.initialize(URI.create(namenodeURI), conf); | ||
return fs; | ||
} catch (IOException ioe) { | ||
throw new RuntimeException(ioe); | ||
} | ||
} | ||
}); | ||
fsMap.put(proxyUser, fs); | ||
|
||
// Load default path | ||
String fileParentPath = conf.get(FILE_PARENT_PATH_KEY, FILE_PARENT_PATH_DEFAULT); | ||
parentFolder = new Path(new Path(fileParentPath), "mapper" + taskID); | ||
|
||
// Make job folder | ||
fs.mkdirs(parentFolder); | ||
|
||
setupLargeListingJob(fs); | ||
if (taskID > numLargeJobs && !enableWriteOps) { | ||
setupSmallListingJob(fs, taskID); | ||
} | ||
} | ||
|
||
private void setupLargeListingJob(FileSystem fs) throws IOException { | ||
for (int i = 0; i < numFiles; i++) { | ||
fs.mkdirs(new Path(parentFolder, new Path("big", new Path("sub" + i)))); | ||
} | ||
} | ||
|
||
private void setupSmallListingJob(FileSystem fs, int taskID) throws IOException { | ||
fs.mkdirs(new Path(parentFolder, new Path("small" + taskID))); | ||
} | ||
|
||
@Override | ||
public void map(LongWritable key, NullWritable value, Context mapperContext) | ||
throws IOException { | ||
int taskID = mapperContext.getTaskAttemptID().getTaskID().getId(); | ||
String proxyUser = "fakeuser" + taskID / mappersPerUser; | ||
FileSystem fs = fsMap.get(proxyUser); | ||
|
||
if (taskID < numLargeJobs) { | ||
runLargeListingJob(fs); | ||
} else if (enableWriteOps) { | ||
runSmallMkdirJob(fs, taskID); | ||
} else { | ||
runSmallListingJob(fs, taskID); | ||
} | ||
} | ||
|
||
private void runLargeListingJob(FileSystem fs) throws IOException { | ||
fs.listStatus(new Path(parentFolder, new Path("big"))); | ||
} | ||
|
||
private void runSmallMkdirJob(FileSystem fs, int taskID) throws IOException { | ||
fs.mkdirs(new Path(parentFolder, new Path("small" + taskID))); | ||
fs.delete(new Path(parentFolder, new Path("small" + taskID)), true); | ||
} | ||
|
||
private void runSmallListingJob(FileSystem fs, int taskID) throws IOException { | ||
fs.listStatus(new Path(parentFolder, new Path("small" + taskID))); | ||
} | ||
} |
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I feel like the wording is a bit weird on some of these descriptions, suggestions welcome!