problem with extraction of .tgz file on Redhat AS 3.0

Girish N girish.narasanna at oracle.com
Mon Dec 20 15:57:56 UTC 2004


Hi Donald,

Yes, the database is shutting down perfectly without any errors before 
the backup starts up, hence no file is in use during backup.

Thanks & Regards
Girish

O'Neill, Donald (US - Deerfield) wrote:

>There is a good possibility. I assumed that your shutting down Oracle
>before your performing the backup? 
>
>-----Original Message-----
>From: redhat-list-bounces at redhat.com
>[mailto:redhat-list-bounces at redhat.com] On Behalf Of Girish N
>Sent: Monday, December 20, 2004 8:18 AM
>To: General Red Hat Linux discussion list
>Subject: Re: problem with extraction of .tgz file on Redhat AS 3.0
>
>Hi Donald,
>
>Thanks for the reply, does this mean that .tgz will get corrupted if the
>
>datafile is in use?
>
>Thx & Rgds
>Girish
>
>O'Neill, Donald (US - Deerfield) wrote:
>
>  
>
>>Some Oracle process could still have some open files. I would do a
>>'fuser -am' on the partition just to make sure there are no open files
>>before tarring.. You could also do a 'tar -cjvf' which uses bzip
>>    
>>
>instead
>  
>
>>of gzip..  
>>
>>-----Original Message-----
>>From: redhat-list-bounces at redhat.com
>>[mailto:redhat-list-bounces at redhat.com] On Behalf Of Girish N
>>Sent: Monday, December 20, 2004 1:56 AM
>>To: General Red Hat Linux discussion list
>>Subject: Re: problem with extraction of .tgz file on Redhat AS 3.0
>>
>>Hi Linus,
>>
>>Thanks again for the reply. As said earlier, wil reschedule the .tgz 
>>dump to a local mount point & will check the same.
>>
>>Thanks & Regards
>>Girish
>>
>>C. Linus Hicks wrote:
>>
>> 
>>
>>    
>>
>>>On Mon, 2004-12-20 at 11:07 +0530, Girish N wrote:
>>>
>>>
>>>   
>>>
>>>      
>>>
>>>>Hi Linus,
>>>>
>>>>Thanks for the reply,
>>>>
>>>>1. The datafile in question if only 1Gb
>>>>2. This is a low end server with 2Gb memory & the backup is scheduled
>>>>     
>>>>
>>>>        
>>>>
>>to 
>> 
>>
>>    
>>
>>>>run at 4 AM when there is no memory resource crunch.
>>>>
>>>>The corruption seems to be very inconsistent, 1 day the .tgz is fine,
>>>>        
>>>>
>
>  
>
>>>>while the 2nd day, the .tgz file gets corrupted. Am planning to 
>>>>reschedule the .tgz backup to one of the local Mount points instead
>>>>        
>>>>
>of
>  
>
>>>>     
>>>>
>>>>        
>>>>
>> 
>>
>>    
>>
>>>>the SAN Harddisk & then check the same.
>>>>
>>>>You hv commented that "Not with the symptoms you have", does that
>>>>        
>>>>
>mean
>  
>
>>>>     
>>>>
>>>>        
>>>>
>> 
>>
>>    
>>
>>>>that this may be one-of-the reasons for file corruption.
>>>>  
>>>>
>>>>     
>>>>
>>>>        
>>>>
>>>Having an inconsistent datafile will not cause the kind of corruption
>>>you are getting in the tgz file. If you backup (by whatever means) a
>>>datafile that is in an inconsistent state, then the result of
>>>      
>>>
>restoring
>  
>
>>>that file will be a datafile in an inconsistent state, not a problem
>>>with the restore. The reason tar complained was because of the gzip
>>>error. When gzip took the error, it was unable to continue ungzipping
>>>the file and sent EOF to tar.
>>>
>>>This means the error will be with corruption either during gzip, or
>>>writing to disk. This suggests a hardware problem, perhaps in memory,
>>>   
>>>
>>>      
>>>
>>or
>> 
>>
>>    
>>
>>>with writing to the SAN. Trying a local disk rather than the SAN is a
>>>good idea. You might also try running memtest on this machine. Having
>>>   
>>>
>>>      
>>>
>>no
>> 
>>
>>    
>>
>>>memory resource crunch at the time of the backup doesn't really mean
>>>much, but I would expect other files to show the same symptom if
>>>      
>>>
>memory
>  
>
>>>is the problem.
>>>
>>>http://www.memtest86.com/
>>>
>>>
>>>
>>>   
>>>
>>>      
>>>



More information about the redhat-list mailing list