You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-user@hadoop.apache.org by Ulul <ha...@ulul.org> on 2014/10/06 23:22:04 UTC
Re: Reduce fails always
Hello
Did you check you don't have a job.setNumReduceTasks(1); in your job
driver ?
And you should check the number of slots available on the jobtracker web
interface
Ulul
Le 06/10/2014 20:34, Abdul Navaz a écrit :
> Hello,
>
> I have 8 Datanodes and each having storage capacity of only 3GB. I am
> running word count on 1GB of text file.
>
> Initially df –h shows it has 2.8GB after HDFS write. When Shuffling
> Starts it goes on consuming the disc space of only one node. I think
> it is the reducer. Finally df –h shows 2MB. Why can’t it just use all
> 4 reducer disc space ?
>
>
>
> Thanks & Regards,
>
> Abdul Navaz
> Research Assistant
> University of Houston Main Campus, Houston TX
> Ph: 281-685-0388
>
>
> From: Rohith Sharma K S <rohithsharmaks@huawei.com
> <ma...@huawei.com>>
> Reply-To: <user@hadoop.apache.org <ma...@hadoop.apache.org>>
> Date: Monday, October 6, 2014 at 5:52 AM
> To: "user@hadoop.apache.org <ma...@hadoop.apache.org>"
> <user@hadoop.apache.org <ma...@hadoop.apache.org>>
> Subject: RE: Reduce fails always
>
> Hi
>
> How much data does wordcount job is processing?
>
> What is the disk space (“df -h” ) available in the node where it
> always fail?
>
> The point I didn’t understand is why it uses only one datanode disc space?
>
> >> For reducers task running, containers can be allocated at any node. I think, in your cluster one of the
> machines disk space is very low. So whichever the task running on that
> particular node is failing.
>
> Thanks & Regards
>
> Rohith Sharma K S
>
> *From:*Abdul Navaz [mailto:navaz.enc@gmail.com]
> *Sent:* 06 October 2014 08:21
> *To:* user@hadoop.apache.org <ma...@hadoop.apache.org>
> *Subject:* Reduce fails always
>
> Hi All,
>
> I am running sample word count job in a 9 node cluster and I am
> getting the below error message.
>
> hadoop jar chiu-wordcount2.jar WordCount /user/hduser/getty/file1.txt
> /user/hduser/getty/out10 -D mapred.reduce.tasks=2
>
> 14/10/05 18:08:45 INFO mapred.JobClient: map 99% reduce 26%
>
> 14/10/05 18:08:48 INFO mapred.JobClient: map 99% reduce 28%
>
> 14/10/05 18:08:51 INFO mapred.JobClient: map 100% reduce 28%
>
> 14/10/05 18:08:57 INFO mapred.JobClient: map 98% reduce 0%
>
> 14/10/05 18:08:58 INFO mapred.JobClient: Task Id :
> attempt_201410051754_0003_r_000000_0, Status : FAILED
>
> FSError: java.io.IOException: No space left on device
>
> 14/10/05 18:08:59 WARN mapred.JobClient: Error reading task
> outputhttp://pcvm1-10.utahddc.geniracks.net:50060/tasklog?plaintext=true&attemptid=attempt_201410051754_0003_r_000000_0&filter=stdout
>
> 14/10/05 18:08:59 WARN mapred.JobClient: Error reading task
> outputhttp://pcvm1-10.utahddc.geniracks.net:50060/tasklog?plaintext=true&attemptid=attempt_201410051754_0003_r_000000_0&filter=stderr
>
> 14/10/05 18:08:59 INFO mapred.JobClient: Task Id :
> attempt_201410051754_0003_m_000015_0, Status : FAILED
>
> FSError: java.io.IOException: No space left on device
>
> 14/10/05 18:09:02 INFO mapred.JobClient: map 99% reduce 0%
>
> 14/10/05 18:09:07 INFO mapred.JobClient: map 99% reduce 1%
>
> I can see it uses all disk space on one of the datanode when shuffling
> starts. As soon as disc space on the node becomes nill it throws me
> this error and job aborts. The point I didn’t understand is why it
> uses only one datanode disc space. I have change the number of
> reducer as 4 still it uses only one datanode disc and throws above error.
>
> How can I fix this issue?
>
> Thanks & Regards,
>
> Navaz
>
Re: Reduce fails always
Posted by Abdul Navaz <na...@gmail.com>.
It was the space issue. It works perfectly after adding this to mapped.site
xml file.
<property>
<name>mapred.local.dir</name>
<value>/mnt</value>
</property>
Thanks for the help. :)
Thanks & Regards,
Abdul Navaz
Research Assistant
University of Houston Main Campus, Houston TX
Ph: 281-685-0388
From: Ulul <ha...@ulul.org>
Reply-To: <us...@hadoop.apache.org>
Date: Monday, October 6, 2014 at 4:22 PM
To: <us...@hadoop.apache.org>
Subject: Re: Reduce fails always
Hello
Did you check you don't have a job.setNumReduceTasks(1); in your job driver
?
And you should check the number of slots available on the jobtracker web
interface
Ulul
Le 06/10/2014 20:34, Abdul Navaz a écrit :
>
>
> Hello,
>
>
>
>
> I have 8 Datanodes and each having storage capacity of only 3GB. I am running
> word count on 1GB of text file.
>
>
>
>
> Initially df h shows it has 2.8GB after HDFS write. When Shuffling Starts it
> goes on consuming the disc space of only one node. I think it is the reducer.
> Finally df h shows 2MB. Why can¹t it just use all 4 reducer disc space ?
>
>
>
>
>
>
>
>
>
>
>
> Thanks & Regards,
>
>
>
>
> Abdul Navaz
>
> Research Assistant
>
> University of Houston Main Campus, Houston TX
>
> Ph: 281-685-0388
>
>
>
>
>
>
>
>
>
> From: Rohith Sharma K S <ro...@huawei.com>
> Reply-To: <us...@hadoop.apache.org>
> Date: Monday, October 6, 2014 at 5:52 AM
> To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
> Subject: RE: Reduce fails always
>
>
>
>
>
>
>
>
>
> Hi
>
>
>
> How much data does wordcount job is processing?
>
> What is the disk space (³df -h² ) available in the node where it always fail?
>
>
>
> The point I didn¹t understand is why it uses only one datanode disc space?
>
>>> >> For reducers task running, containers can be allocated at any node. I
>>> think, in your cluster one of the machines disk space is very low. So
>>> whichever the task running on that particular node is failing.
>
>
>
>
>
> Thanks & Regards
>
> Rohith Sharma K S
>
>
>
>
>
>
>
>
> From: Abdul Navaz [mailto:navaz.enc@gmail.com]
> Sent: 06 October 2014 08:21
> To: user@hadoop.apache.org
> Subject: Reduce fails always
>
>
>
>
>
>
>
> Hi All,
>
>
>
>
>
>
>
>
>
> I am running sample word count job in a 9 node cluster and I am getting the
> below error message.
>
>
>
>
>
>
>
>
>
> hadoop jar chiu-wordcount2.jar WordCount /user/hduser/getty/file1.txt
> /user/hduser/getty/out10 -D mapred.reduce.tasks=2
>
>
>
>
>
>
>
>
>
> 14/10/05 18:08:45 INFO mapred.JobClient: map 99% reduce 26%
>
>
>
>
> 14/10/05 18:08:48 INFO mapred.JobClient: map 99% reduce 28%
>
>
> 14/10/05 18:08:51 INFO mapred.JobClient: map 100% reduce 28%
>
>
> 14/10/05 18:08:57 INFO mapred.JobClient: map 98% reduce 0%
>
>
> 14/10/05 18:08:58 INFO mapred.JobClient: Task Id :
> attempt_201410051754_0003_r_000000_0, Status : FAILED
>
>
> FSError: java.io.IOException: No space left on device
>
>
> 14/10/05 18:08:59 WARN mapred.JobClient: Error reading task
> outputhttp://pcvm1-10.utahddc.geniracks.net:50060/tasklog?plaintext=true&attem
> ptid=attempt_201410051754_0003_r_000000_0&filter=stdout
>
>
> 14/10/05 18:08:59 WARN mapred.JobClient: Error reading task
> outputhttp://pcvm1-10.utahddc.geniracks.net:50060/tasklog?plaintext=true&attem
> ptid=attempt_201410051754_0003_r_000000_0&filter=stderr
>
>
> 14/10/05 18:08:59 INFO mapred.JobClient: Task Id :
> attempt_201410051754_0003_m_000015_0, Status : FAILED
>
>
> FSError: java.io.IOException: No space left on device
>
>
> 14/10/05 18:09:02 INFO mapred.JobClient: map 99% reduce 0%
>
>
> 14/10/05 18:09:07 INFO mapred.JobClient: map 99% reduce 1%
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> I can see it uses all disk space on one of the datanode when shuffling starts.
> As soon as disc space on the node becomes nill it throws me this error and job
> aborts. The point I didn¹t understand is why it uses only one datanode disc
> space. I have change the number of reducer as 4 still it uses only one
> datanode disc and throws above error.
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> How can I fix this issue?
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> Thanks & Regards,
>
>
>
>
>
>
>
>
>
> Navaz
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
Re: Reduce fails always
Posted by Abdul Navaz <na...@gmail.com>.
It was the space issue. It works perfectly after adding this to mapped.site
xml file.
<property>
<name>mapred.local.dir</name>
<value>/mnt</value>
</property>
Thanks for the help. :)
Thanks & Regards,
Abdul Navaz
Research Assistant
University of Houston Main Campus, Houston TX
Ph: 281-685-0388
From: Ulul <ha...@ulul.org>
Reply-To: <us...@hadoop.apache.org>
Date: Monday, October 6, 2014 at 4:22 PM
To: <us...@hadoop.apache.org>
Subject: Re: Reduce fails always
Hello
Did you check you don't have a job.setNumReduceTasks(1); in your job driver
?
And you should check the number of slots available on the jobtracker web
interface
Ulul
Le 06/10/2014 20:34, Abdul Navaz a écrit :
>
>
> Hello,
>
>
>
>
> I have 8 Datanodes and each having storage capacity of only 3GB. I am running
> word count on 1GB of text file.
>
>
>
>
> Initially df h shows it has 2.8GB after HDFS write. When Shuffling Starts it
> goes on consuming the disc space of only one node. I think it is the reducer.
> Finally df h shows 2MB. Why can¹t it just use all 4 reducer disc space ?
>
>
>
>
>
>
>
>
>
>
>
> Thanks & Regards,
>
>
>
>
> Abdul Navaz
>
> Research Assistant
>
> University of Houston Main Campus, Houston TX
>
> Ph: 281-685-0388
>
>
>
>
>
>
>
>
>
> From: Rohith Sharma K S <ro...@huawei.com>
> Reply-To: <us...@hadoop.apache.org>
> Date: Monday, October 6, 2014 at 5:52 AM
> To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
> Subject: RE: Reduce fails always
>
>
>
>
>
>
>
>
>
> Hi
>
>
>
> How much data does wordcount job is processing?
>
> What is the disk space (³df -h² ) available in the node where it always fail?
>
>
>
> The point I didn¹t understand is why it uses only one datanode disc space?
>
>>> >> For reducers task running, containers can be allocated at any node. I
>>> think, in your cluster one of the machines disk space is very low. So
>>> whichever the task running on that particular node is failing.
>
>
>
>
>
> Thanks & Regards
>
> Rohith Sharma K S
>
>
>
>
>
>
>
>
> From: Abdul Navaz [mailto:navaz.enc@gmail.com]
> Sent: 06 October 2014 08:21
> To: user@hadoop.apache.org
> Subject: Reduce fails always
>
>
>
>
>
>
>
> Hi All,
>
>
>
>
>
>
>
>
>
> I am running sample word count job in a 9 node cluster and I am getting the
> below error message.
>
>
>
>
>
>
>
>
>
> hadoop jar chiu-wordcount2.jar WordCount /user/hduser/getty/file1.txt
> /user/hduser/getty/out10 -D mapred.reduce.tasks=2
>
>
>
>
>
>
>
>
>
> 14/10/05 18:08:45 INFO mapred.JobClient: map 99% reduce 26%
>
>
>
>
> 14/10/05 18:08:48 INFO mapred.JobClient: map 99% reduce 28%
>
>
> 14/10/05 18:08:51 INFO mapred.JobClient: map 100% reduce 28%
>
>
> 14/10/05 18:08:57 INFO mapred.JobClient: map 98% reduce 0%
>
>
> 14/10/05 18:08:58 INFO mapred.JobClient: Task Id :
> attempt_201410051754_0003_r_000000_0, Status : FAILED
>
>
> FSError: java.io.IOException: No space left on device
>
>
> 14/10/05 18:08:59 WARN mapred.JobClient: Error reading task
> outputhttp://pcvm1-10.utahddc.geniracks.net:50060/tasklog?plaintext=true&attem
> ptid=attempt_201410051754_0003_r_000000_0&filter=stdout
>
>
> 14/10/05 18:08:59 WARN mapred.JobClient: Error reading task
> outputhttp://pcvm1-10.utahddc.geniracks.net:50060/tasklog?plaintext=true&attem
> ptid=attempt_201410051754_0003_r_000000_0&filter=stderr
>
>
> 14/10/05 18:08:59 INFO mapred.JobClient: Task Id :
> attempt_201410051754_0003_m_000015_0, Status : FAILED
>
>
> FSError: java.io.IOException: No space left on device
>
>
> 14/10/05 18:09:02 INFO mapred.JobClient: map 99% reduce 0%
>
>
> 14/10/05 18:09:07 INFO mapred.JobClient: map 99% reduce 1%
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> I can see it uses all disk space on one of the datanode when shuffling starts.
> As soon as disc space on the node becomes nill it throws me this error and job
> aborts. The point I didn¹t understand is why it uses only one datanode disc
> space. I have change the number of reducer as 4 still it uses only one
> datanode disc and throws above error.
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> How can I fix this issue?
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> Thanks & Regards,
>
>
>
>
>
>
>
>
>
> Navaz
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
Re: Reduce fails always
Posted by Abdul Navaz <na...@gmail.com>.
It was the space issue. It works perfectly after adding this to mapped.site
xml file.
<property>
<name>mapred.local.dir</name>
<value>/mnt</value>
</property>
Thanks for the help. :)
Thanks & Regards,
Abdul Navaz
Research Assistant
University of Houston Main Campus, Houston TX
Ph: 281-685-0388
From: Ulul <ha...@ulul.org>
Reply-To: <us...@hadoop.apache.org>
Date: Monday, October 6, 2014 at 4:22 PM
To: <us...@hadoop.apache.org>
Subject: Re: Reduce fails always
Hello
Did you check you don't have a job.setNumReduceTasks(1); in your job driver
?
And you should check the number of slots available on the jobtracker web
interface
Ulul
Le 06/10/2014 20:34, Abdul Navaz a écrit :
>
>
> Hello,
>
>
>
>
> I have 8 Datanodes and each having storage capacity of only 3GB. I am running
> word count on 1GB of text file.
>
>
>
>
> Initially df h shows it has 2.8GB after HDFS write. When Shuffling Starts it
> goes on consuming the disc space of only one node. I think it is the reducer.
> Finally df h shows 2MB. Why can¹t it just use all 4 reducer disc space ?
>
>
>
>
>
>
>
>
>
>
>
> Thanks & Regards,
>
>
>
>
> Abdul Navaz
>
> Research Assistant
>
> University of Houston Main Campus, Houston TX
>
> Ph: 281-685-0388
>
>
>
>
>
>
>
>
>
> From: Rohith Sharma K S <ro...@huawei.com>
> Reply-To: <us...@hadoop.apache.org>
> Date: Monday, October 6, 2014 at 5:52 AM
> To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
> Subject: RE: Reduce fails always
>
>
>
>
>
>
>
>
>
> Hi
>
>
>
> How much data does wordcount job is processing?
>
> What is the disk space (³df -h² ) available in the node where it always fail?
>
>
>
> The point I didn¹t understand is why it uses only one datanode disc space?
>
>>> >> For reducers task running, containers can be allocated at any node. I
>>> think, in your cluster one of the machines disk space is very low. So
>>> whichever the task running on that particular node is failing.
>
>
>
>
>
> Thanks & Regards
>
> Rohith Sharma K S
>
>
>
>
>
>
>
>
> From: Abdul Navaz [mailto:navaz.enc@gmail.com]
> Sent: 06 October 2014 08:21
> To: user@hadoop.apache.org
> Subject: Reduce fails always
>
>
>
>
>
>
>
> Hi All,
>
>
>
>
>
>
>
>
>
> I am running sample word count job in a 9 node cluster and I am getting the
> below error message.
>
>
>
>
>
>
>
>
>
> hadoop jar chiu-wordcount2.jar WordCount /user/hduser/getty/file1.txt
> /user/hduser/getty/out10 -D mapred.reduce.tasks=2
>
>
>
>
>
>
>
>
>
> 14/10/05 18:08:45 INFO mapred.JobClient: map 99% reduce 26%
>
>
>
>
> 14/10/05 18:08:48 INFO mapred.JobClient: map 99% reduce 28%
>
>
> 14/10/05 18:08:51 INFO mapred.JobClient: map 100% reduce 28%
>
>
> 14/10/05 18:08:57 INFO mapred.JobClient: map 98% reduce 0%
>
>
> 14/10/05 18:08:58 INFO mapred.JobClient: Task Id :
> attempt_201410051754_0003_r_000000_0, Status : FAILED
>
>
> FSError: java.io.IOException: No space left on device
>
>
> 14/10/05 18:08:59 WARN mapred.JobClient: Error reading task
> outputhttp://pcvm1-10.utahddc.geniracks.net:50060/tasklog?plaintext=true&attem
> ptid=attempt_201410051754_0003_r_000000_0&filter=stdout
>
>
> 14/10/05 18:08:59 WARN mapred.JobClient: Error reading task
> outputhttp://pcvm1-10.utahddc.geniracks.net:50060/tasklog?plaintext=true&attem
> ptid=attempt_201410051754_0003_r_000000_0&filter=stderr
>
>
> 14/10/05 18:08:59 INFO mapred.JobClient: Task Id :
> attempt_201410051754_0003_m_000015_0, Status : FAILED
>
>
> FSError: java.io.IOException: No space left on device
>
>
> 14/10/05 18:09:02 INFO mapred.JobClient: map 99% reduce 0%
>
>
> 14/10/05 18:09:07 INFO mapred.JobClient: map 99% reduce 1%
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> I can see it uses all disk space on one of the datanode when shuffling starts.
> As soon as disc space on the node becomes nill it throws me this error and job
> aborts. The point I didn¹t understand is why it uses only one datanode disc
> space. I have change the number of reducer as 4 still it uses only one
> datanode disc and throws above error.
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> How can I fix this issue?
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> Thanks & Regards,
>
>
>
>
>
>
>
>
>
> Navaz
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
Re: Reduce fails always
Posted by Abdul Navaz <na...@gmail.com>.
It was the space issue. It works perfectly after adding this to mapped.site
xml file.
<property>
<name>mapred.local.dir</name>
<value>/mnt</value>
</property>
Thanks for the help. :)
Thanks & Regards,
Abdul Navaz
Research Assistant
University of Houston Main Campus, Houston TX
Ph: 281-685-0388
From: Ulul <ha...@ulul.org>
Reply-To: <us...@hadoop.apache.org>
Date: Monday, October 6, 2014 at 4:22 PM
To: <us...@hadoop.apache.org>
Subject: Re: Reduce fails always
Hello
Did you check you don't have a job.setNumReduceTasks(1); in your job driver
?
And you should check the number of slots available on the jobtracker web
interface
Ulul
Le 06/10/2014 20:34, Abdul Navaz a écrit :
>
>
> Hello,
>
>
>
>
> I have 8 Datanodes and each having storage capacity of only 3GB. I am running
> word count on 1GB of text file.
>
>
>
>
> Initially df h shows it has 2.8GB after HDFS write. When Shuffling Starts it
> goes on consuming the disc space of only one node. I think it is the reducer.
> Finally df h shows 2MB. Why can¹t it just use all 4 reducer disc space ?
>
>
>
>
>
>
>
>
>
>
>
> Thanks & Regards,
>
>
>
>
> Abdul Navaz
>
> Research Assistant
>
> University of Houston Main Campus, Houston TX
>
> Ph: 281-685-0388
>
>
>
>
>
>
>
>
>
> From: Rohith Sharma K S <ro...@huawei.com>
> Reply-To: <us...@hadoop.apache.org>
> Date: Monday, October 6, 2014 at 5:52 AM
> To: "user@hadoop.apache.org" <us...@hadoop.apache.org>
> Subject: RE: Reduce fails always
>
>
>
>
>
>
>
>
>
> Hi
>
>
>
> How much data does wordcount job is processing?
>
> What is the disk space (³df -h² ) available in the node where it always fail?
>
>
>
> The point I didn¹t understand is why it uses only one datanode disc space?
>
>>> >> For reducers task running, containers can be allocated at any node. I
>>> think, in your cluster one of the machines disk space is very low. So
>>> whichever the task running on that particular node is failing.
>
>
>
>
>
> Thanks & Regards
>
> Rohith Sharma K S
>
>
>
>
>
>
>
>
> From: Abdul Navaz [mailto:navaz.enc@gmail.com]
> Sent: 06 October 2014 08:21
> To: user@hadoop.apache.org
> Subject: Reduce fails always
>
>
>
>
>
>
>
> Hi All,
>
>
>
>
>
>
>
>
>
> I am running sample word count job in a 9 node cluster and I am getting the
> below error message.
>
>
>
>
>
>
>
>
>
> hadoop jar chiu-wordcount2.jar WordCount /user/hduser/getty/file1.txt
> /user/hduser/getty/out10 -D mapred.reduce.tasks=2
>
>
>
>
>
>
>
>
>
> 14/10/05 18:08:45 INFO mapred.JobClient: map 99% reduce 26%
>
>
>
>
> 14/10/05 18:08:48 INFO mapred.JobClient: map 99% reduce 28%
>
>
> 14/10/05 18:08:51 INFO mapred.JobClient: map 100% reduce 28%
>
>
> 14/10/05 18:08:57 INFO mapred.JobClient: map 98% reduce 0%
>
>
> 14/10/05 18:08:58 INFO mapred.JobClient: Task Id :
> attempt_201410051754_0003_r_000000_0, Status : FAILED
>
>
> FSError: java.io.IOException: No space left on device
>
>
> 14/10/05 18:08:59 WARN mapred.JobClient: Error reading task
> outputhttp://pcvm1-10.utahddc.geniracks.net:50060/tasklog?plaintext=true&attem
> ptid=attempt_201410051754_0003_r_000000_0&filter=stdout
>
>
> 14/10/05 18:08:59 WARN mapred.JobClient: Error reading task
> outputhttp://pcvm1-10.utahddc.geniracks.net:50060/tasklog?plaintext=true&attem
> ptid=attempt_201410051754_0003_r_000000_0&filter=stderr
>
>
> 14/10/05 18:08:59 INFO mapred.JobClient: Task Id :
> attempt_201410051754_0003_m_000015_0, Status : FAILED
>
>
> FSError: java.io.IOException: No space left on device
>
>
> 14/10/05 18:09:02 INFO mapred.JobClient: map 99% reduce 0%
>
>
> 14/10/05 18:09:07 INFO mapred.JobClient: map 99% reduce 1%
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> I can see it uses all disk space on one of the datanode when shuffling starts.
> As soon as disc space on the node becomes nill it throws me this error and job
> aborts. The point I didn¹t understand is why it uses only one datanode disc
> space. I have change the number of reducer as 4 still it uses only one
> datanode disc and throws above error.
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> How can I fix this issue?
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> Thanks & Regards,
>
>
>
>
>
>
>
>
>
> Navaz
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>