I found an application that interacts with one of googles APIs that sends requests to the API and returns results to the web page. The owner of this app has made it publicly available and even though it is hosted on his server, it acts directly with the api from anyones browser. The source code in page source is available but since I dont know javascript, it is meaningless to me.
I would like to create an app that calls this web app and have it send requests to the API, and have the results parsed to a database. Would this be possible?
I have very little ASP experience and since trying to work through accessing the API itself is complicated since it involves using oAuth protocol and I am having a hard time with this with not much sample code available.
If this is possible, how can I get access to the results from this web app?
I have a dataGrid with ButtonColumns in my asp page and I would like to do the following to:
When a row is clicked, I want to be able to:
1 - Hide all rows and display only the clicked row dynamically/on the fly.
trying to loop thru the datagrid and display only the Row index passed by the OnITemComman - but unsure of how to do it.
2 - Extract the row values (all cells in the row) to use as parameters in a query.
Used many ways of many examples but i always get a blank value. Managed to get the button text in the class but i need the full row values not just the cell clicked.
I have a datalist that is populated via and SqlDataSource. Code is VB
I am trying to extract two Values from the Item Templete and add them together. Then I want to add a new Label to the item template with the sum of the two added values.
I am able to to this using a GridView but no success with the datalist
I am developing a simple online DBMS to browse my database online.... the main concern for me is to clear everything from the PC when the user disconnect so no body can go and click (Back) on the browser and see the pages if the user forgot to close it.. what I have done so far is :
[Code]....
I am sure this is not enough.. am I right? so what else I should include there?
I want to create a Online Exam site in ASP.NET .In that i want to display Test based on the Date and time and the user can take the test at the time set my the Teacher.
For Example:
Teacher want to take C Test for the Mid Semester Test and the test is schedule on 1 May 2010,then the test should appear on the test page on the 1st may only and the user should be able to take the test on the set time.
The problem is that if i use System time then user can change the system time and can take the test schedule in future.
So i want to know how can implement this thing.I have given many online test but dont know how they work out the time and date logic.
This may be something simple, but I am completely new to ASP.NET. I have been reading some tutorials based on W3Schools, and I have copied one of the files and tried to play with it by changing an element with a value to storing that same value in the attribute. Here is the edited XML (my.xml):
Is there a reliable way to extract text from PDF? The first thought that comes to mind is that PDF may have multiple columns and the extraction mechanism needs to know the logical structure somehow. I understand that some PDF docs are "tagged" but I'd need to support pretty much any PDF document.
public void ZipExtract(Stream inputStream, string outputDirectory) { using (ZipFile zip = ZipFile.Read(inputStream)) { Directory.CreateDirectory(outputDirectory); zip.ExtractSelectedEntries("name=*.jpg,*.jpeg,*.png,*.gif,*.bmp", " ", outputDirectory, ExtractExistingFileAction.OverwriteSilently);
[Code]....
I try to read zip archive from stream and extract files. Got the following exception in the line "using (ZipFile zip = ZipFile.Read(inputStream))" : ZipEntry::ReadDirEntry(): Bad signature (0xC618F879) at position 0x0000EE19. how to handle this exception?
Every morning we get a few hundred emails from our various customers servers that tell us how their backup went that previous night. Somebody has to go through these and fill out an excel spread sheet. I would like to automate this, we have our own server running a .NET application so i can schedule something every morning.
There are two main parts. The first is i want to know how to create a table that has customers down the left and dates accross the right, they are called pivot tables in access but not sure how to recreate them in MSSQL and to do it on the fly as i dont want to have to sit there entering 100's of days at a time.
The second is that differnt customers use differnt backup solutions, so i want to write some code that scans a particular mailbox in exchange, looks for emails that have come in in the past 12 hours, look at the from email address and match that to one in the database, then according to the subject line it will be able to identify what type of backup solution it is, e.g. if it says "Backup Exec Job completd Sucessfully" it knows its backup exec and will look for a particular text (completed successfully, completed with exceptions, no media etc) and can insert that, or if its a differnet backup solution then go and look at a specific place in that email for it.
<A HREF="/files/1234/mybook.b1">mybook.b1</A><br>11/24/2010 1:02 AM 991
I want to fetch only the date element from this string. i.e 11/24/2010 1:02 AM
I know i can do that using the substring method of string class but i need to implement a standard solution for this problem. i.e. fetch date from any html content if it contains a date element.
I have a requirement wherein a SQL query will be given in a text file.
The text file has to be read , the SQL query needs to be parsed and the column names, table names and where clauses have to be extracted and inserted into separate tables.
i am using the below code for extracting paragraph for matching string.
[code]...
i am getting correct result,but i am getting breakable words in starting and ending paragraphs like "...ing regions uses the and Boolean connector to specify the region so narr..."
I have a Portable Bridge Notation formatted-file that I have to work with. I have already a few simple examples working by using indexing and substrings extracting what I need and I suppose for this PBN business it would suite me well since it isn't run too often. If I one the other hand would run that code like all the time (thinking of working with vCards) then I am worried about memory usage under high workload because of the high amout of temporary string variables created from all substrings and splits.
There are two other options available that I know of. Regex and StringReader / TextReader and I wanted some general opinion on what to use.
Intended usage is to extract to objects and serialize to json so that I can more easily work with or persist this information. Hell if it's fast enough I might even do it on the fly.
The task is: given a web application that returns a table with pagination, do a software that "reads and parses it" since there is nothing like a webservice that provides the raw data. It's like a "spider" or a "crawler" application to steal data that is not meant to be accessed programmatically.
Now the thing: the application is made with standart aspx webform engine, so nothing like standard URLs or posts, but the dreadful postback engine crowded with javascript and non accessible html. The pagination links call the infamous javascript:__doPostBack(param, param) so I think it wouldn't even work if I try even to simulate clicks on those links.
There are also inputs to filter the results and they are also part of the postback mechanism, so I can't simulate a regular post to get the results.
I was forced to do something like this in the past, but it was on a standard-like website with parameters in the querystring like pagesize and pagenumber so I was able to sort it out.
Anyone has a vague idea if this is doable, or if I should tell to my boss to quit asking me to do this retarded stuff?
EDIT: maybe I was a bit unclear about what I have to achieve. I have to parse, extract and convert that data in another format - let's say excel - and not just read it. And this stuff must be automated without user input. I don't think Selenium would cut it.
EDIT: I just blogged about this situation. If anyone is interested can check my post at [URL] and comment about that.
I'm reading an excel file using oledb in my asp.net code (vb).I have no problem going over the data but in one of the cells there is a string that "hides" a hyperlink. I want that hyperlink and not the string...
I referenced this page, how to extract data from a gridview on a webpage. [URL] ....
I can extract to a Word, Excel and CSV but I can't to a pdf file. I get the following error:
Server Error in '/' Application.
Security Exception
Description: The application attempted to perform an operation not allowed by the security policy. To grant this application the required permission please contact your system administrator or change the application's trust level in the configuration file. Exception Details: System.Security.SecurityException: That assembly does not allow partially trusted callers.Source Error:
Line 135: Response.Write(pdfDoc); Line 136: Response.End(); Line 137: } Line 138: Line 139: protected void btnExportCSV_Click(object sender, EventArgs e)
As you know viewstate can recognize as a vulnarable item if we do not encrype it I know how to encrype the viewstate that I define in a web page but my problem start from here I downloaded a software which help you to exract the unencrypted viewstate of every website the link of that software is as follow:
[URL]
but when I make even empty page without viewstate ,this program can extract the viewstae of my page but if check some website like google ,yahoo,amazoon
I have a textbox that gets the date from an AJAX Calendar extender. What I need to do is extract the "month," "day," and "year," and put them into seperate string variables to use in my code to do various things based on the month, day or year respectively.
Here is what I have so far:
[code]....
But I don't want the year NOW, I want the year from the textbox.
I've used the substring feature before, but I am struggling with this one since I only want the page names which are on the right hand side, I basically want to chop of everything else (folder paths).. all I want to end up with is default.aspx, gallery.aspx and demo.aspx
I have almost 100 website that will update in a condition, I have a winzip archive that contains the files that replaces those websites. I want to know that
I can extract that files in a folder and then copy them to all 100 websites folders
I can extract the archive directly to 100 websites folders
which one is better in performance and less prone to errors