Corrupt PDF files When File is written to the response.

E

eventuranza

For my application, there users can upload and download files to the a
webserver. Straightforward enough. However, when they upload a PDF
file then try to download it, the file seems to be corrupted. The had
not been a problem before but it seems to have been introduced when I
refactored the writing to the response to a buffered version to prevent
potential OutOfMemoryExceptions (you wouldn't believe the size of some
of these files...).

Here's the code (pretty much jacked from a Microsoft knowledge base
article):

private void WriteToResponse(string filepath, string contentType,
NameValueCollection extraHeaders)
{
Stream iStream = null;

// Buffer to read 10K bytes in chunk:
byte[] buffer = new Byte[10000];

// Length of the file:
int length;

// Total bytes to read:
long dataToRead;

// Identify the file name.
string filename = Path.GetFileName(filepath);

try
{
// Open the file.
iStream = new FileStream(filepath, FileMode.Open,
FileAccess.Read, FileShare.Read);


// Total bytes to read:
dataToRead = iStream.Length;

Response.ContentType = contentType;
if(extraHeaders != null)
{
for (int i = 0; i < extraHeaders.Count; i++)
{
Trace.Write("adding Key value: " +
extraHeaders.GetKey(i)
+ ". Adding value: " +
extraHeaders);
Response.AddHeader(extraHeaders.GetKey(i),
extraHeaders);
}
}

// Read the bytes.
while (dataToRead > 0)
{
// Verify that the client is connected.
if (Response.IsClientConnected)
{
// Read the data in buffer.
length = iStream.Read(buffer, 0, 10000);

// Write the data to the current output stream.
Response.OutputStream.Write(buffer, 0, length);

// Flush the data to the HTML output.
Response.Flush();

buffer = new Byte[10000];
dataToRead = dataToRead - length;
}
else
{
//prevent infinite loop if user disconnects
dataToRead = -1;
}
}
}
catch(IOException ioEx)
{
errorRow.Visible = true;
lblErrorMsg.Text = ioEx.Message;
log.Error(ioEx.ToString());
}
catch (Exception ex)
{
errorRow.Visible = true;
// Trap the error, if any.
log.Error(ex.ToString());
}
finally
{
if (iStream != null)
{
//Close the file.
iStream.Close();
}
}
}

Thanks in advance for any help.
 
P

Patrice

Trace not enabled ? All HTML stripped ? (using Response.Clear allows to make
sure you won't have a character coming from your APSX markup).

IMO The easiest path is to save the file (possibly using a content
disposition header) so that you can check the length. If not the length you
can do a file comparison to see where the file is corrupted.
 
T

Tim_Mac

hi,
did you try using Response.BufferOutput = true;
i thought that was supposed to buffer the output to the client in smaller
chunks.
..Net 2.0 has a method called Response.TransmitFile() which according to the
docs says that it sends a file straight to the client without loading the
contents into memory. it must do a direct stream transfer from the file to
the http response.

the thought of doing tracing in the middle of a binary write makes me a bit
suspicious, surely this would corrupt the download? not sure though if the
trace strings actually get written out with the response.

thanks
tim

For my application, there users can upload and download files to the a
webserver. Straightforward enough. However, when they upload a PDF
file then try to download it, the file seems to be corrupted. The had
not been a problem before but it seems to have been introduced when I
refactored the writing to the response to a buffered version to prevent
potential OutOfMemoryExceptions (you wouldn't believe the size of some
of these files...).

Here's the code (pretty much jacked from a Microsoft knowledge base
article):

private void WriteToResponse(string filepath, string contentType,
NameValueCollection extraHeaders)
{
Stream iStream = null;

// Buffer to read 10K bytes in chunk:
byte[] buffer = new Byte[10000];

// Length of the file:
int length;

// Total bytes to read:
long dataToRead;

// Identify the file name.
string filename = Path.GetFileName(filepath);

try
{
// Open the file.
iStream = new FileStream(filepath, FileMode.Open,
FileAccess.Read, FileShare.Read);


// Total bytes to read:
dataToRead = iStream.Length;

Response.ContentType = contentType;
if(extraHeaders != null)
{
for (int i = 0; i < extraHeaders.Count; i++)
{
Trace.Write("adding Key value: " +
extraHeaders.GetKey(i)
+ ". Adding value: " +
extraHeaders);
Response.AddHeader(extraHeaders.GetKey(i),
extraHeaders);
}
}

// Read the bytes.
while (dataToRead > 0)
{
// Verify that the client is connected.
if (Response.IsClientConnected)
{
// Read the data in buffer.
length = iStream.Read(buffer, 0, 10000);

// Write the data to the current output stream.
Response.OutputStream.Write(buffer, 0, length);

// Flush the data to the HTML output.
Response.Flush();

buffer = new Byte[10000];
dataToRead = dataToRead - length;
}
else
{
//prevent infinite loop if user disconnects
dataToRead = -1;
}
}
}
catch(IOException ioEx)
{
errorRow.Visible = true;
lblErrorMsg.Text = ioEx.Message;
log.Error(ioEx.ToString());
}
catch (Exception ex)
{
errorRow.Visible = true;
// Trap the error, if any.
log.Error(ex.ToString());
}
finally
{
if (iStream != null)
{
//Close the file.
iStream.Close();
}
}
}

Thanks in advance for any help.
 
E

eventuranza

Thanks for all your input. What it turned to be was my neglecting to
put in "Response.End()" at the end of this method. The files that were
being downloaded actually had the emitted HTML response appended to the
end of the file. It turns out that out of out of a whole cross section
of files, ranging form movie files to Office and images, the PDF was
most sensitive to this extra information after the EOF and was gettin
gcorrupted.


Tim_Mac said:
hi,
did you try using Response.BufferOutput = true;
i thought that was supposed to buffer the output to the client in smaller
chunks.
.Net 2.0 has a method called Response.TransmitFile() which according to the
docs says that it sends a file straight to the client without loading the
contents into memory. it must do a direct stream transfer from the file to
the http response.

the thought of doing tracing in the middle of a binary write makes me a bit
suspicious, surely this would corrupt the download? not sure though if the
trace strings actually get written out with the response.

thanks
tim

For my application, there users can upload and download files to the a
webserver. Straightforward enough. However, when they upload a PDF
file then try to download it, the file seems to be corrupted. The had
not been a problem before but it seems to have been introduced when I
refactored the writing to the response to a buffered version to prevent
potential OutOfMemoryExceptions (you wouldn't believe the size of some
of these files...).

Here's the code (pretty much jacked from a Microsoft knowledge base
article):

private void WriteToResponse(string filepath, string contentType,
NameValueCollection extraHeaders)
{
Stream iStream = null;

// Buffer to read 10K bytes in chunk:
byte[] buffer = new Byte[10000];

// Length of the file:
int length;

// Total bytes to read:
long dataToRead;

// Identify the file name.
string filename = Path.GetFileName(filepath);

try
{
// Open the file.
iStream = new FileStream(filepath, FileMode.Open,
FileAccess.Read, FileShare.Read);


// Total bytes to read:
dataToRead = iStream.Length;

Response.ContentType = contentType;
if(extraHeaders != null)
{
for (int i = 0; i < extraHeaders.Count; i++)
{
Trace.Write("adding Key value: " +
extraHeaders.GetKey(i)
+ ". Adding value: " +
extraHeaders);
Response.AddHeader(extraHeaders.GetKey(i),
extraHeaders);
}
}

// Read the bytes.
while (dataToRead > 0)
{
// Verify that the client is connected.
if (Response.IsClientConnected)
{
// Read the data in buffer.
length = iStream.Read(buffer, 0, 10000);

// Write the data to the current output stream.
Response.OutputStream.Write(buffer, 0, length);

// Flush the data to the HTML output.
Response.Flush();

buffer = new Byte[10000];
dataToRead = dataToRead - length;
}
else
{
//prevent infinite loop if user disconnects
dataToRead = -1;
}
}
}
catch(IOException ioEx)
{
errorRow.Visible = true;
lblErrorMsg.Text = ioEx.Message;
log.Error(ioEx.ToString());
}
catch (Exception ex)
{
errorRow.Visible = true;
// Trap the error, if any.
log.Error(ex.ToString());
}
finally
{
if (iStream != null)
{
//Close the file.
iStream.Close();
}
}
}

Thanks in advance for any help.
 
D

Damien

Thanks for all your input. What it turned to be was my neglecting to
put in "Response.End()" at the end of this method. The files that were
being downloaded actually had the emitted HTML response appended to the
end of the file. It turns out that out of out of a whole cross section
of files, ranging form movie files to Office and images, the PDF was
most sensitive to this extra information after the EOF and was gettin
gcorrupted.
Glad you sorted it, but can I just ask, why do you continually
reallocate the buffer in your loop? Surely that would increase the
memory pressure?

Damien
 
T

Tim_Mac

hi,
you may prefer to use Response.Close()
in .net 1.1 i ran into ThreadAbortExceptions when using Response.End.
changing this to Response.Close() solved the problem.

tim
 

Ask a Question

Want to reply to this thread or ask your own question?

You'll need to choose a username for the site, which only take a couple of moments. After that, you can post your question and our members will help you out.

Ask a Question

Members online

No members online now.

Forum statistics

Threads
473,744
Messages
2,569,484
Members
44,903
Latest member
orderPeak8CBDGummies

Latest Threads

Top