You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@hadoop.apache.org by "Venkivolu, Dayakar Reddy" <Da...@ca.com> on 2013/06/08 04:49:56 UTC
hdfsConnect/hdfsWrite API writes conetnts of file to local system
instead of HDFS system
Hi,
I have created the sample program to write the contents into HDFS file system. The file gets created successfully, but unfortunately the file is getting created in
Local system instead of HDFS system.
Here is the source code of sample program:
int main(int argc, char **argv) {
const char* writePath = "/user/testuser/test1.txt";
const char *tuser = "root";
hdfsFS fs = NULL;
int exists = 0;
fs = hdfsConnectAsUser("default", 0, tuser);
if(fs == NULL ) {
fprintf(stderr, "Oops! Failed to connect to hdfs!\n");
exit(-1);
}
hdfsFile writeFile = hdfsOpenFile(fs, writePath, O_WRONLY|O_CREAT, 0, 0, 0);
if(!writeFile) {
fprintf(stderr, "Failed to open %s for writing!\n", writePath);
exit(-1);
}
fprintf(stderr, "Opened %s for writing successfully...\n", writePath);
char* buffer = "Hello, World!";
tSize num_written_bytes = hdfsWrite(fs, writeFile, (void*)buffer, strlen(buffer)+1);
fprintf(stderr, "Wrote %d bytes\n", num_written_bytes);
fprintf(stderr, "Flushed %s successfully!\n", writePath);
hdfsCloseFile(fs, writeFile);
}
CLASSPATH
/usr/lib/hadoop/lib/activation-1.1.jar:/usr/lib/hadoop/lib/asm-3.2.jar:/usr/lib/hadoop/lib/avro-1.7.3.jar:/usr/lib/hadoop/lib/commons-beanutils-1.7.0.jar:
/usr/lib/hadoop/lib/commons-beanutils-core-1.8.0.jar:/usr/lib/hadoop/lib/commons-cli-1.2.jar:/usr/lib/hadoop/lib/commons-codec-1.4.jar:
/usr/lib/hadoop/lib/commons-collections-3.2.1.jar:/usr/lib/hadoop/lib/commons-configuration-1.6.jar:/usr/lib/hadoop/lib/commons-digester-1.8.jar:
/usr/lib/hadoop/lib/commons-el-1.0.jar:/usr/lib/hadoop/lib/commons-httpclient-3.1.jar:/usr/lib/hadoop/lib/commons-io-2.1.jar:/usr/lib/hadoop/lib/commons-lang-2.5.jar:
/usr/lib/hadoop/lib/commons-logging-1.1.1.jar:/usr/lib/hadoop/lib/commons-math-2.1.jar:/usr/lib/hadoop/lib/commons-net-3.1.jar:/usr/lib/hadoop/lib/guava-11.0.2.jar:
/usr/lib/hadoop/lib/hue-plugins-2.2.0-cdh4.2.0.jar:/usr/lib/hadoop/lib/jackson-core-asl-1.8.8.jar:/usr/lib/hadoop/lib/jackson-jaxrs-1.8.8.jar:
/usr/lib/hadoop/lib/jackson-mapper-asl-1.8.8.jar:/usr/lib/hadoop/lib/jackson-xc-1.8.8.jar:/usr/lib/hadoop/lib/jackson-xc-1.8.8.jar:
/usr/lib/hadoop/lib/jasper-compiler-5.5.23.jar:/usr/lib/hadoop/lib/jasper-runtime-5.5.23.jar:/usr/lib/hadoop/lib/jaxb-api-2.2.2.jar:/usr/lib/hadoop/lib/jaxb-impl-2.2.3-1.jar:
/usr/lib/hadoop/lib/jersey-core-1.8.jar:/usr/lib/hadoop/lib/jersey-json-1.8.jar:/usr/lib/hadoop/lib/jersey-server-1.8.jar:/usr/lib/hadoop/lib/jets3t-0.6.1.jar:/usr/lib/hadoop/lib/jettison-1.1.jar:
/usr/lib/hadoop/lib/jetty-6.1.26.cloudera.2.jar: /usr/lib/hadoop/lib/jetty-util-6.1.26.cloudera.2.jar:/usr/lib/hadoop/lib/jline-0.9.94.jar:/usr/lib/hadoop/lib/jsch-0.1.42.jar:
/usr/lib/hadoop/lib/jsp-api-2.1.jar:/usr/lib/hadoop/lib/jsr305-1.3.9.jar:/usr/lib/hadoop/lib/junit-4.8.2.jar:/usr/lib/hadoop/lib/kfs-0.3.jar:/usr/lib/hadoop/lib/log4j-1.2.17.jar:
/usr/lib/hadoop/lib/mockito-all-1.8.5.jar:/usr/lib/hadoop/lib/paranamer-2.3.jar:/usr/lib/hadoop/lib/protobuf-java-2.4.0a.jar:/usr/lib/hadoop/lib/servlet-api-2.5.jar:
/usr/lib/hadoop/lib/slf4j-api-1.6.1.jar:/usr/lib/hadoop/lib/slf4j-log4j12-1.6.1.jar:/usr/lib/hadoop/lib/snappy-java-1.0.4.1.jar:/usr/lib/hadoop/lib/stax-api-1.0.1.jar:
/usr/lib/hadoop/lib/xmlenc-0.52.jar:/usr/lib/hadoop/lib/zookeeper-3.4.5-cdh4.2.0.jar:/usr/lib/hadoop/hadoop-annotations.jar:/usr/lib/hadoop/hadoop-auth-2.0.0-cdh4.2.0.jar:
/usr/lib/hadoop/hadoop-auth.jar:/usr/lib/hadoop/hadoop-common-2.0.0-hdfs_site.xml<http://hadoop.6.n7.nabble.com/file/n69039/hdfs_site.xml>core-site.xml<http://hadoop.6.n7.nabble.com/file/n69039/core-site.xml>cdh4.2.0.jar:/usr/lib/hadoop/hadoop-common-2.0.0-cdh4.2.0-tests.jar:/usr/lib/hadoop/hadoop-common.jar:
/usr/lib/hadoop/etc/hadoop/yarn-site.xml:/usr/lib/hadoop/etc/hadoop/core-site.xml:/usr/lib/hadoop/etc/hadoop/hadoop-metrics.properties:/usr/lib/hadoop/etc/hadoop/hdfs-site.xml:
/usr/lib/hadoop/etc/hadoop/mapred-site.xml
Please find attached the hdfs_site.xml and core_site.xml.
Regards,
Dayakar
Re: hdfsConnect/hdfsWrite API writes conetnts of file to local system
instead of HDFS system
Posted by Shahab Yunus <sh...@gmail.com>.
I notice that for fs.default.name, you have in core-site.xml
hdfs://127.0.0.1:8020
This is local host address, of your local machine so hdfs is there. Where
you are expecting it to be created and where it is being created?fs
Regards,
Shahab
On Fri, Jun 7, 2013 at 10:49 PM, Venkivolu, Dayakar Reddy <
DayakarReddy.Venkivolu@ca.com> wrote:
> Hi,****
>
> ** **
>
> I have created the sample program to write the contents into HDFS file
> system. The file gets created successfully, but unfortunately the file is
> getting created in****
>
> Local system instead of HDFS system.****
>
> ** **
>
> Here is the source code of sample program:****
>
> ** **
>
> *int main(int argc, char **argv) {*
>
> * *
>
> * const char* writePath = "/user/testuser/test1.txt";*
>
> * const char *tuser = "root";*
>
> * hdfsFS fs = NULL;*
>
> * int exists = 0;*
>
> * *
>
> * fs = hdfsConnectAsUser("default", 0, tuser);*
>
> * if(fs == NULL ) {*
>
> * fprintf(stderr, "Oops! Failed to connect to hdfs!\n");*
>
> * exit(-1);*
>
> * }*
>
> * *
>
> * hdfsFile writeFile = hdfsOpenFile(fs, writePath,
> O_WRONLY|O_CREAT, 0, 0, 0);*
>
> * if(!writeFile) {*
>
> * fprintf(stderr, "Failed to open %s for
> writing!\n", writePath);*
>
> * exit(-1);*
>
> * }*
>
> * *
>
> * fprintf(stderr, "Opened %s for writing
> successfully...\n", writePath);*
>
> * *
>
> * char* buffer = "Hello, World!";*
>
> * tSize num_written_bytes = hdfsWrite(fs, writeFile,
> (void*)buffer, strlen(buffer)+1);*
>
> * fprintf(stderr, "Wrote %d bytes\n", num_written_bytes);
> *
>
> * *
>
> * fprintf(stderr, "Flushed %s successfully!\n",
> writePath);*
>
> * hdfsCloseFile(fs, writeFile);*
>
> } ****
>
> ** **
>
> *CLASSPATH*
>
> /usr/lib/hadoop/lib/activation-1.1.jar:/usr/lib/hadoop/lib/asm-3.2.jar:/usr/lib/hadoop/lib/avro-1.7.3.jar:/usr/lib/hadoop/lib/commons-beanutils-1.7.0.jar:
>
>
> /usr/lib/hadoop/lib/commons-beanutils-core-1.8.0.jar:/usr/lib/hadoop/lib/commons-cli-1.2.jar:/usr/lib/hadoop/lib/commons-codec-1.4.jar:
>
>
> /usr/lib/hadoop/lib/commons-collections-3.2.1.jar:/usr/lib/hadoop/lib/commons-configuration-1.6.jar:/usr/lib/hadoop/lib/commons-digester-1.8.jar:
>
>
> /usr/lib/hadoop/lib/commons-el-1.0.jar:/usr/lib/hadoop/lib/commons-httpclient-3.1.jar:/usr/lib/hadoop/lib/commons-io-2.1.jar:/usr/lib/hadoop/lib/commons-lang-2.5.jar:
>
>
> /usr/lib/hadoop/lib/commons-logging-1.1.1.jar:/usr/lib/hadoop/lib/commons-math-2.1.jar:/usr/lib/hadoop/lib/commons-net-3.1.jar:/usr/lib/hadoop/lib/guava-11.0.2.jar:
>
>
> /usr/lib/hadoop/lib/hue-plugins-2.2.0-cdh4.2.0.jar:/usr/lib/hadoop/lib/jackson-core-asl-1.8.8.jar:/usr/lib/hadoop/lib/jackson-jaxrs-1.8.8.jar:
>
>
> /usr/lib/hadoop/lib/jackson-mapper-asl-1.8.8.jar:/usr/lib/hadoop/lib/jackson-xc-1.8.8.jar:/usr/lib/hadoop/lib/jackson-xc-1.8.8.jar:
>
>
> /usr/lib/hadoop/lib/jasper-compiler-5.5.23.jar:/usr/lib/hadoop/lib/jasper-runtime-5.5.23.jar:/usr/lib/hadoop/lib/jaxb-api-2.2.2.jar:/usr/lib/hadoop/lib/jaxb-impl-2.2.3-1.jar:
>
>
> /usr/lib/hadoop/lib/jersey-core-1.8.jar:/usr/lib/hadoop/lib/jersey-json-1.8.jar:/usr/lib/hadoop/lib/jersey-server-1.8.jar:/usr/lib/hadoop/lib/jets3t-0.6.1.jar:/usr/lib/hadoop/lib/jettison-1.1.jar:
>
> /usr/lib/hadoop/lib/jetty-6.1.26.cloudera.2.jar:
> /usr/lib/hadoop/lib/jetty-util-6.1.26.cloudera.2.jar:/usr/lib/hadoop/lib/jline-0.9.94.jar:/usr/lib/hadoop/lib/jsch-0.1.42.jar:
>
>
> /usr/lib/hadoop/lib/jsp-api-2.1.jar:/usr/lib/hadoop/lib/jsr305-1.3.9.jar:/usr/lib/hadoop/lib/junit-4.8.2.jar:/usr/lib/hadoop/lib/kfs-0.3.jar:/usr/lib/hadoop/lib/log4j-1.2.17.jar:
>
>
> /usr/lib/hadoop/lib/mockito-all-1.8.5.jar:/usr/lib/hadoop/lib/paranamer-2.3.jar:/usr/lib/hadoop/lib/protobuf-java-2.4.0a.jar:/usr/lib/hadoop/lib/servlet-api-2.5.jar:
>
>
> /usr/lib/hadoop/lib/slf4j-api-1.6.1.jar:/usr/lib/hadoop/lib/slf4j-log4j12-1.6.1.jar:/usr/lib/hadoop/lib/snappy-java-1.0.4.1.jar:/usr/lib/hadoop/lib/stax-api-1.0.1.jar:
>
>
> /usr/lib/hadoop/lib/xmlenc-0.52.jar:/usr/lib/hadoop/lib/zookeeper-3.4.5-cdh4.2.0.jar:/usr/lib/hadoop/hadoop-annotations.jar:/usr/lib/hadoop/hadoop-auth-2.0.0-cdh4.2.0.jar:
>
>
> /usr/lib/hadoop/hadoop-auth.jar:/usr/lib/hadoop/hadoop-common-2.0.0-
> hdfs_site.xml <http://hadoop.6.n7.nabble.com/file/n69039/hdfs_site.xml>
> core-site.xml <http://hadoop.6.n7.nabble.com/file/n69039/core-site.xml>cdh4.2.0.jar:/usr/lib/hadoop/hadoop-common-2.0.0-cdh4.2.0-tests.jar:/usr/lib/hadoop/hadoop-common.jar:
>
>
> /usr/lib/hadoop/etc/hadoop/yarn-site.xml:/usr/lib/hadoop/etc/hadoop/core-site.xml:/usr/lib/hadoop/etc/hadoop/hadoop-metrics.properties:/usr/lib/hadoop/etc/hadoop/hdfs-site.xml:
>
> /usr/lib/hadoop/etc/hadoop/mapred-site.xml****
>
> ** **
>
> Please find attached the hdfs_site.xml and core_site.xml.****
>
> ** **
>
> Regards,****
>
> Dayakar****
>
Re: hdfsConnect/hdfsWrite API writes conetnts of file to local system
instead of HDFS system
Posted by Shahab Yunus <sh...@gmail.com>.
I notice that for fs.default.name, you have in core-site.xml
hdfs://127.0.0.1:8020
This is local host address, of your local machine so hdfs is there. Where
you are expecting it to be created and where it is being created?fs
Regards,
Shahab
On Fri, Jun 7, 2013 at 10:49 PM, Venkivolu, Dayakar Reddy <
DayakarReddy.Venkivolu@ca.com> wrote:
> Hi,****
>
> ** **
>
> I have created the sample program to write the contents into HDFS file
> system. The file gets created successfully, but unfortunately the file is
> getting created in****
>
> Local system instead of HDFS system.****
>
> ** **
>
> Here is the source code of sample program:****
>
> ** **
>
> *int main(int argc, char **argv) {*
>
> * *
>
> * const char* writePath = "/user/testuser/test1.txt";*
>
> * const char *tuser = "root";*
>
> * hdfsFS fs = NULL;*
>
> * int exists = 0;*
>
> * *
>
> * fs = hdfsConnectAsUser("default", 0, tuser);*
>
> * if(fs == NULL ) {*
>
> * fprintf(stderr, "Oops! Failed to connect to hdfs!\n");*
>
> * exit(-1);*
>
> * }*
>
> * *
>
> * hdfsFile writeFile = hdfsOpenFile(fs, writePath,
> O_WRONLY|O_CREAT, 0, 0, 0);*
>
> * if(!writeFile) {*
>
> * fprintf(stderr, "Failed to open %s for
> writing!\n", writePath);*
>
> * exit(-1);*
>
> * }*
>
> * *
>
> * fprintf(stderr, "Opened %s for writing
> successfully...\n", writePath);*
>
> * *
>
> * char* buffer = "Hello, World!";*
>
> * tSize num_written_bytes = hdfsWrite(fs, writeFile,
> (void*)buffer, strlen(buffer)+1);*
>
> * fprintf(stderr, "Wrote %d bytes\n", num_written_bytes);
> *
>
> * *
>
> * fprintf(stderr, "Flushed %s successfully!\n",
> writePath);*
>
> * hdfsCloseFile(fs, writeFile);*
>
> } ****
>
> ** **
>
> *CLASSPATH*
>
> /usr/lib/hadoop/lib/activation-1.1.jar:/usr/lib/hadoop/lib/asm-3.2.jar:/usr/lib/hadoop/lib/avro-1.7.3.jar:/usr/lib/hadoop/lib/commons-beanutils-1.7.0.jar:
>
>
> /usr/lib/hadoop/lib/commons-beanutils-core-1.8.0.jar:/usr/lib/hadoop/lib/commons-cli-1.2.jar:/usr/lib/hadoop/lib/commons-codec-1.4.jar:
>
>
> /usr/lib/hadoop/lib/commons-collections-3.2.1.jar:/usr/lib/hadoop/lib/commons-configuration-1.6.jar:/usr/lib/hadoop/lib/commons-digester-1.8.jar:
>
>
> /usr/lib/hadoop/lib/commons-el-1.0.jar:/usr/lib/hadoop/lib/commons-httpclient-3.1.jar:/usr/lib/hadoop/lib/commons-io-2.1.jar:/usr/lib/hadoop/lib/commons-lang-2.5.jar:
>
>
> /usr/lib/hadoop/lib/commons-logging-1.1.1.jar:/usr/lib/hadoop/lib/commons-math-2.1.jar:/usr/lib/hadoop/lib/commons-net-3.1.jar:/usr/lib/hadoop/lib/guava-11.0.2.jar:
>
>
> /usr/lib/hadoop/lib/hue-plugins-2.2.0-cdh4.2.0.jar:/usr/lib/hadoop/lib/jackson-core-asl-1.8.8.jar:/usr/lib/hadoop/lib/jackson-jaxrs-1.8.8.jar:
>
>
> /usr/lib/hadoop/lib/jackson-mapper-asl-1.8.8.jar:/usr/lib/hadoop/lib/jackson-xc-1.8.8.jar:/usr/lib/hadoop/lib/jackson-xc-1.8.8.jar:
>
>
> /usr/lib/hadoop/lib/jasper-compiler-5.5.23.jar:/usr/lib/hadoop/lib/jasper-runtime-5.5.23.jar:/usr/lib/hadoop/lib/jaxb-api-2.2.2.jar:/usr/lib/hadoop/lib/jaxb-impl-2.2.3-1.jar:
>
>
> /usr/lib/hadoop/lib/jersey-core-1.8.jar:/usr/lib/hadoop/lib/jersey-json-1.8.jar:/usr/lib/hadoop/lib/jersey-server-1.8.jar:/usr/lib/hadoop/lib/jets3t-0.6.1.jar:/usr/lib/hadoop/lib/jettison-1.1.jar:
>
> /usr/lib/hadoop/lib/jetty-6.1.26.cloudera.2.jar:
> /usr/lib/hadoop/lib/jetty-util-6.1.26.cloudera.2.jar:/usr/lib/hadoop/lib/jline-0.9.94.jar:/usr/lib/hadoop/lib/jsch-0.1.42.jar:
>
>
> /usr/lib/hadoop/lib/jsp-api-2.1.jar:/usr/lib/hadoop/lib/jsr305-1.3.9.jar:/usr/lib/hadoop/lib/junit-4.8.2.jar:/usr/lib/hadoop/lib/kfs-0.3.jar:/usr/lib/hadoop/lib/log4j-1.2.17.jar:
>
>
> /usr/lib/hadoop/lib/mockito-all-1.8.5.jar:/usr/lib/hadoop/lib/paranamer-2.3.jar:/usr/lib/hadoop/lib/protobuf-java-2.4.0a.jar:/usr/lib/hadoop/lib/servlet-api-2.5.jar:
>
>
> /usr/lib/hadoop/lib/slf4j-api-1.6.1.jar:/usr/lib/hadoop/lib/slf4j-log4j12-1.6.1.jar:/usr/lib/hadoop/lib/snappy-java-1.0.4.1.jar:/usr/lib/hadoop/lib/stax-api-1.0.1.jar:
>
>
> /usr/lib/hadoop/lib/xmlenc-0.52.jar:/usr/lib/hadoop/lib/zookeeper-3.4.5-cdh4.2.0.jar:/usr/lib/hadoop/hadoop-annotations.jar:/usr/lib/hadoop/hadoop-auth-2.0.0-cdh4.2.0.jar:
>
>
> /usr/lib/hadoop/hadoop-auth.jar:/usr/lib/hadoop/hadoop-common-2.0.0-
> hdfs_site.xml <http://hadoop.6.n7.nabble.com/file/n69039/hdfs_site.xml>
> core-site.xml <http://hadoop.6.n7.nabble.com/file/n69039/core-site.xml>cdh4.2.0.jar:/usr/lib/hadoop/hadoop-common-2.0.0-cdh4.2.0-tests.jar:/usr/lib/hadoop/hadoop-common.jar:
>
>
> /usr/lib/hadoop/etc/hadoop/yarn-site.xml:/usr/lib/hadoop/etc/hadoop/core-site.xml:/usr/lib/hadoop/etc/hadoop/hadoop-metrics.properties:/usr/lib/hadoop/etc/hadoop/hdfs-site.xml:
>
> /usr/lib/hadoop/etc/hadoop/mapred-site.xml****
>
> ** **
>
> Please find attached the hdfs_site.xml and core_site.xml.****
>
> ** **
>
> Regards,****
>
> Dayakar****
>
Re: hdfsConnect/hdfsWrite API writes conetnts of file to local system
instead of HDFS system
Posted by Shahab Yunus <sh...@gmail.com>.
I notice that for fs.default.name, you have in core-site.xml
hdfs://127.0.0.1:8020
This is local host address, of your local machine so hdfs is there. Where
you are expecting it to be created and where it is being created?fs
Regards,
Shahab
On Fri, Jun 7, 2013 at 10:49 PM, Venkivolu, Dayakar Reddy <
DayakarReddy.Venkivolu@ca.com> wrote:
> Hi,****
>
> ** **
>
> I have created the sample program to write the contents into HDFS file
> system. The file gets created successfully, but unfortunately the file is
> getting created in****
>
> Local system instead of HDFS system.****
>
> ** **
>
> Here is the source code of sample program:****
>
> ** **
>
> *int main(int argc, char **argv) {*
>
> * *
>
> * const char* writePath = "/user/testuser/test1.txt";*
>
> * const char *tuser = "root";*
>
> * hdfsFS fs = NULL;*
>
> * int exists = 0;*
>
> * *
>
> * fs = hdfsConnectAsUser("default", 0, tuser);*
>
> * if(fs == NULL ) {*
>
> * fprintf(stderr, "Oops! Failed to connect to hdfs!\n");*
>
> * exit(-1);*
>
> * }*
>
> * *
>
> * hdfsFile writeFile = hdfsOpenFile(fs, writePath,
> O_WRONLY|O_CREAT, 0, 0, 0);*
>
> * if(!writeFile) {*
>
> * fprintf(stderr, "Failed to open %s for
> writing!\n", writePath);*
>
> * exit(-1);*
>
> * }*
>
> * *
>
> * fprintf(stderr, "Opened %s for writing
> successfully...\n", writePath);*
>
> * *
>
> * char* buffer = "Hello, World!";*
>
> * tSize num_written_bytes = hdfsWrite(fs, writeFile,
> (void*)buffer, strlen(buffer)+1);*
>
> * fprintf(stderr, "Wrote %d bytes\n", num_written_bytes);
> *
>
> * *
>
> * fprintf(stderr, "Flushed %s successfully!\n",
> writePath);*
>
> * hdfsCloseFile(fs, writeFile);*
>
> } ****
>
> ** **
>
> *CLASSPATH*
>
> /usr/lib/hadoop/lib/activation-1.1.jar:/usr/lib/hadoop/lib/asm-3.2.jar:/usr/lib/hadoop/lib/avro-1.7.3.jar:/usr/lib/hadoop/lib/commons-beanutils-1.7.0.jar:
>
>
> /usr/lib/hadoop/lib/commons-beanutils-core-1.8.0.jar:/usr/lib/hadoop/lib/commons-cli-1.2.jar:/usr/lib/hadoop/lib/commons-codec-1.4.jar:
>
>
> /usr/lib/hadoop/lib/commons-collections-3.2.1.jar:/usr/lib/hadoop/lib/commons-configuration-1.6.jar:/usr/lib/hadoop/lib/commons-digester-1.8.jar:
>
>
> /usr/lib/hadoop/lib/commons-el-1.0.jar:/usr/lib/hadoop/lib/commons-httpclient-3.1.jar:/usr/lib/hadoop/lib/commons-io-2.1.jar:/usr/lib/hadoop/lib/commons-lang-2.5.jar:
>
>
> /usr/lib/hadoop/lib/commons-logging-1.1.1.jar:/usr/lib/hadoop/lib/commons-math-2.1.jar:/usr/lib/hadoop/lib/commons-net-3.1.jar:/usr/lib/hadoop/lib/guava-11.0.2.jar:
>
>
> /usr/lib/hadoop/lib/hue-plugins-2.2.0-cdh4.2.0.jar:/usr/lib/hadoop/lib/jackson-core-asl-1.8.8.jar:/usr/lib/hadoop/lib/jackson-jaxrs-1.8.8.jar:
>
>
> /usr/lib/hadoop/lib/jackson-mapper-asl-1.8.8.jar:/usr/lib/hadoop/lib/jackson-xc-1.8.8.jar:/usr/lib/hadoop/lib/jackson-xc-1.8.8.jar:
>
>
> /usr/lib/hadoop/lib/jasper-compiler-5.5.23.jar:/usr/lib/hadoop/lib/jasper-runtime-5.5.23.jar:/usr/lib/hadoop/lib/jaxb-api-2.2.2.jar:/usr/lib/hadoop/lib/jaxb-impl-2.2.3-1.jar:
>
>
> /usr/lib/hadoop/lib/jersey-core-1.8.jar:/usr/lib/hadoop/lib/jersey-json-1.8.jar:/usr/lib/hadoop/lib/jersey-server-1.8.jar:/usr/lib/hadoop/lib/jets3t-0.6.1.jar:/usr/lib/hadoop/lib/jettison-1.1.jar:
>
> /usr/lib/hadoop/lib/jetty-6.1.26.cloudera.2.jar:
> /usr/lib/hadoop/lib/jetty-util-6.1.26.cloudera.2.jar:/usr/lib/hadoop/lib/jline-0.9.94.jar:/usr/lib/hadoop/lib/jsch-0.1.42.jar:
>
>
> /usr/lib/hadoop/lib/jsp-api-2.1.jar:/usr/lib/hadoop/lib/jsr305-1.3.9.jar:/usr/lib/hadoop/lib/junit-4.8.2.jar:/usr/lib/hadoop/lib/kfs-0.3.jar:/usr/lib/hadoop/lib/log4j-1.2.17.jar:
>
>
> /usr/lib/hadoop/lib/mockito-all-1.8.5.jar:/usr/lib/hadoop/lib/paranamer-2.3.jar:/usr/lib/hadoop/lib/protobuf-java-2.4.0a.jar:/usr/lib/hadoop/lib/servlet-api-2.5.jar:
>
>
> /usr/lib/hadoop/lib/slf4j-api-1.6.1.jar:/usr/lib/hadoop/lib/slf4j-log4j12-1.6.1.jar:/usr/lib/hadoop/lib/snappy-java-1.0.4.1.jar:/usr/lib/hadoop/lib/stax-api-1.0.1.jar:
>
>
> /usr/lib/hadoop/lib/xmlenc-0.52.jar:/usr/lib/hadoop/lib/zookeeper-3.4.5-cdh4.2.0.jar:/usr/lib/hadoop/hadoop-annotations.jar:/usr/lib/hadoop/hadoop-auth-2.0.0-cdh4.2.0.jar:
>
>
> /usr/lib/hadoop/hadoop-auth.jar:/usr/lib/hadoop/hadoop-common-2.0.0-
> hdfs_site.xml <http://hadoop.6.n7.nabble.com/file/n69039/hdfs_site.xml>
> core-site.xml <http://hadoop.6.n7.nabble.com/file/n69039/core-site.xml>cdh4.2.0.jar:/usr/lib/hadoop/hadoop-common-2.0.0-cdh4.2.0-tests.jar:/usr/lib/hadoop/hadoop-common.jar:
>
>
> /usr/lib/hadoop/etc/hadoop/yarn-site.xml:/usr/lib/hadoop/etc/hadoop/core-site.xml:/usr/lib/hadoop/etc/hadoop/hadoop-metrics.properties:/usr/lib/hadoop/etc/hadoop/hdfs-site.xml:
>
> /usr/lib/hadoop/etc/hadoop/mapred-site.xml****
>
> ** **
>
> Please find attached the hdfs_site.xml and core_site.xml.****
>
> ** **
>
> Regards,****
>
> Dayakar****
>
Re: hdfsConnect/hdfsWrite API writes conetnts of file to local system
instead of HDFS system
Posted by Shahab Yunus <sh...@gmail.com>.
I notice that for fs.default.name, you have in core-site.xml
hdfs://127.0.0.1:8020
This is local host address, of your local machine so hdfs is there. Where
you are expecting it to be created and where it is being created?fs
Regards,
Shahab
On Fri, Jun 7, 2013 at 10:49 PM, Venkivolu, Dayakar Reddy <
DayakarReddy.Venkivolu@ca.com> wrote:
> Hi,****
>
> ** **
>
> I have created the sample program to write the contents into HDFS file
> system. The file gets created successfully, but unfortunately the file is
> getting created in****
>
> Local system instead of HDFS system.****
>
> ** **
>
> Here is the source code of sample program:****
>
> ** **
>
> *int main(int argc, char **argv) {*
>
> * *
>
> * const char* writePath = "/user/testuser/test1.txt";*
>
> * const char *tuser = "root";*
>
> * hdfsFS fs = NULL;*
>
> * int exists = 0;*
>
> * *
>
> * fs = hdfsConnectAsUser("default", 0, tuser);*
>
> * if(fs == NULL ) {*
>
> * fprintf(stderr, "Oops! Failed to connect to hdfs!\n");*
>
> * exit(-1);*
>
> * }*
>
> * *
>
> * hdfsFile writeFile = hdfsOpenFile(fs, writePath,
> O_WRONLY|O_CREAT, 0, 0, 0);*
>
> * if(!writeFile) {*
>
> * fprintf(stderr, "Failed to open %s for
> writing!\n", writePath);*
>
> * exit(-1);*
>
> * }*
>
> * *
>
> * fprintf(stderr, "Opened %s for writing
> successfully...\n", writePath);*
>
> * *
>
> * char* buffer = "Hello, World!";*
>
> * tSize num_written_bytes = hdfsWrite(fs, writeFile,
> (void*)buffer, strlen(buffer)+1);*
>
> * fprintf(stderr, "Wrote %d bytes\n", num_written_bytes);
> *
>
> * *
>
> * fprintf(stderr, "Flushed %s successfully!\n",
> writePath);*
>
> * hdfsCloseFile(fs, writeFile);*
>
> } ****
>
> ** **
>
> *CLASSPATH*
>
> /usr/lib/hadoop/lib/activation-1.1.jar:/usr/lib/hadoop/lib/asm-3.2.jar:/usr/lib/hadoop/lib/avro-1.7.3.jar:/usr/lib/hadoop/lib/commons-beanutils-1.7.0.jar:
>
>
> /usr/lib/hadoop/lib/commons-beanutils-core-1.8.0.jar:/usr/lib/hadoop/lib/commons-cli-1.2.jar:/usr/lib/hadoop/lib/commons-codec-1.4.jar:
>
>
> /usr/lib/hadoop/lib/commons-collections-3.2.1.jar:/usr/lib/hadoop/lib/commons-configuration-1.6.jar:/usr/lib/hadoop/lib/commons-digester-1.8.jar:
>
>
> /usr/lib/hadoop/lib/commons-el-1.0.jar:/usr/lib/hadoop/lib/commons-httpclient-3.1.jar:/usr/lib/hadoop/lib/commons-io-2.1.jar:/usr/lib/hadoop/lib/commons-lang-2.5.jar:
>
>
> /usr/lib/hadoop/lib/commons-logging-1.1.1.jar:/usr/lib/hadoop/lib/commons-math-2.1.jar:/usr/lib/hadoop/lib/commons-net-3.1.jar:/usr/lib/hadoop/lib/guava-11.0.2.jar:
>
>
> /usr/lib/hadoop/lib/hue-plugins-2.2.0-cdh4.2.0.jar:/usr/lib/hadoop/lib/jackson-core-asl-1.8.8.jar:/usr/lib/hadoop/lib/jackson-jaxrs-1.8.8.jar:
>
>
> /usr/lib/hadoop/lib/jackson-mapper-asl-1.8.8.jar:/usr/lib/hadoop/lib/jackson-xc-1.8.8.jar:/usr/lib/hadoop/lib/jackson-xc-1.8.8.jar:
>
>
> /usr/lib/hadoop/lib/jasper-compiler-5.5.23.jar:/usr/lib/hadoop/lib/jasper-runtime-5.5.23.jar:/usr/lib/hadoop/lib/jaxb-api-2.2.2.jar:/usr/lib/hadoop/lib/jaxb-impl-2.2.3-1.jar:
>
>
> /usr/lib/hadoop/lib/jersey-core-1.8.jar:/usr/lib/hadoop/lib/jersey-json-1.8.jar:/usr/lib/hadoop/lib/jersey-server-1.8.jar:/usr/lib/hadoop/lib/jets3t-0.6.1.jar:/usr/lib/hadoop/lib/jettison-1.1.jar:
>
> /usr/lib/hadoop/lib/jetty-6.1.26.cloudera.2.jar:
> /usr/lib/hadoop/lib/jetty-util-6.1.26.cloudera.2.jar:/usr/lib/hadoop/lib/jline-0.9.94.jar:/usr/lib/hadoop/lib/jsch-0.1.42.jar:
>
>
> /usr/lib/hadoop/lib/jsp-api-2.1.jar:/usr/lib/hadoop/lib/jsr305-1.3.9.jar:/usr/lib/hadoop/lib/junit-4.8.2.jar:/usr/lib/hadoop/lib/kfs-0.3.jar:/usr/lib/hadoop/lib/log4j-1.2.17.jar:
>
>
> /usr/lib/hadoop/lib/mockito-all-1.8.5.jar:/usr/lib/hadoop/lib/paranamer-2.3.jar:/usr/lib/hadoop/lib/protobuf-java-2.4.0a.jar:/usr/lib/hadoop/lib/servlet-api-2.5.jar:
>
>
> /usr/lib/hadoop/lib/slf4j-api-1.6.1.jar:/usr/lib/hadoop/lib/slf4j-log4j12-1.6.1.jar:/usr/lib/hadoop/lib/snappy-java-1.0.4.1.jar:/usr/lib/hadoop/lib/stax-api-1.0.1.jar:
>
>
> /usr/lib/hadoop/lib/xmlenc-0.52.jar:/usr/lib/hadoop/lib/zookeeper-3.4.5-cdh4.2.0.jar:/usr/lib/hadoop/hadoop-annotations.jar:/usr/lib/hadoop/hadoop-auth-2.0.0-cdh4.2.0.jar:
>
>
> /usr/lib/hadoop/hadoop-auth.jar:/usr/lib/hadoop/hadoop-common-2.0.0-
> hdfs_site.xml <http://hadoop.6.n7.nabble.com/file/n69039/hdfs_site.xml>
> core-site.xml <http://hadoop.6.n7.nabble.com/file/n69039/core-site.xml>cdh4.2.0.jar:/usr/lib/hadoop/hadoop-common-2.0.0-cdh4.2.0-tests.jar:/usr/lib/hadoop/hadoop-common.jar:
>
>
> /usr/lib/hadoop/etc/hadoop/yarn-site.xml:/usr/lib/hadoop/etc/hadoop/core-site.xml:/usr/lib/hadoop/etc/hadoop/hadoop-metrics.properties:/usr/lib/hadoop/etc/hadoop/hdfs-site.xml:
>
> /usr/lib/hadoop/etc/hadoop/mapred-site.xml****
>
> ** **
>
> Please find attached the hdfs_site.xml and core_site.xml.****
>
> ** **
>
> Regards,****
>
> Dayakar****
>