Search Results

Search found 57448 results on 2298 pages for 'microsoft web administrat'.

Page 592/2298 | < Previous Page | 588 589 590 591 592 593 594 595 596 597 598 599  | Next Page >

  • generic async loading method for page web scripts?

    - by boomhauer
    The google analytics code went to an async load model some time back. I've noticed that a lot of the other scripts I use on many sites are causing slow load times - specifically the addthis script and the facebook like button. I'm noticing that the slow load times of these scripts is causing the google bot to calc my page loadtimes as being much slower than previously. I'd like to know if there is a standard/generic way of causing these scripts to load async as well, or perhaps a pointer to someone who has done the work for this already. Seems this would be a popular thing to do, but not much luck searching around.

    Read the article

  • Google Maps: marker icons are multi-drawing over same location, why?

    - by JakiT
    I'm using Google Maps and have the map set to height/width=100%. I noticed that when I resize the browser window, my map marker icons appear to redraw on top of the same location. Meaning, if I should only have 1 map marker icon on "987 Main St", when I resize the browser window - multiple map marker icons are being now draw on top of "987 Main St" that shouldn't be there. Any ideas why this is happening and how to prevent it? UPDATE: Link to actual code

    Read the article

  • image protection in rails

    - by Cezar
    Hello, I am looking for ways to protect my product images and I don't know if there's anything out there better than what I've already found: disable right click, use a transparent image in front of your picture and watermarking. Obviously none of them is perfect but I was curious if someone came up with a better solution to this problem. Also is there any rails plugin to aid with that ? Thanks

    Read the article

  • Breeze Expand not working on WebAPI with EF

    - by Rodney
    I have published a WebAPI service which returns a list of items. I am implementing Breeze and have managed to get it basically working with filtering/sorting. However, the Expand is not working. http://www.ftter.com/desktopmodules/framework/api/dare/dares/?$filter=DareId%20eq%2010&expand%20eq%20ToUser You can see the ToUserId ForeignKey in the response above, but the ToUser properties are NULL (the user definitely exists) You can see the ToUser EF navigation property in the metadata. When I use .Include on the server side I can populate it with EF, but I don't want to do this. [HttpGet] [Queryable(AllowedQueryOptions = AllowedQueryOptions.All)] public HttpResponseMessage Dares() { var response = Request.CreateResponse(HttpStatusCode.OK, (IQueryable<Dare>)contextProvider.Context.Dares); return ControllerUtilities.GetResponseWithCorsHeader(response); } and here is the generated class from my EF model (using Database First) public partial class Dare { public int DareId { get; set; } public int ToUserId { get; set; } public virtual User ToUser { get; set; } }

    Read the article

  • How to make software which will work like torrent?

    - by Nitz
    Hey guys How to make software work like torrent work? Bcz always when i see that torrent software then i am really amazed by their sizes and what they do in that sizes? How they managed the download by parts and then all together as soon as u downloaded full part? and as soon as you had great speed then your download speed automatically goes up? How to make software which will work like torrent means peer-to-peer? how to make this kind of software which can download from different servers and managed to know each users download and upload? and how they have pretty small size? which tech. they have used in that kind of software? Sorry i had asked many question. I know how downloading is happening means peer-to-peer all that.. but i don't know how they have built that kind of thing?

    Read the article

  • Client Web Browser Behavior When Handling 301 Redirect

    - by Jon Swanson
    The RFC seems to suggest that the client should permanently cache the response: http://www.w3.org/Protocols/rfc2616/rfc2616-sec10.html 10.3.2 301 Moved Permanently The requested resource has been assigned a new permanent URI and any future references to this resource SHOULD use one of the returned URIs. Clients with link editing capabilities ought to automatically re-link references to the Request-URI to one or more of the new references returned by the server, where possible. This response is cacheable unless indicated otherwise. The new permanent URI SHOULD be given by the Location field in the response. Unless the request method was HEAD, the entity of the response SHOULD contain a short hypertext note with a hyperlink to the new URI(s). If the 301 status code is received in response to a request other than GET or HEAD, the user agent MUST NOT automatically redirect the request unless it can be confirmed by the user, since this might change the conditions under which the request was issued. Note: When automatically redirecting a POST request after receiving a 301 status code, some existing HTTP/1.0 user agents will erroneously change it into a GET request. I'm having a hard time finding concrete browser documentation for any major browser that states how they handle these. I've started digging through the source code of firefox, but quickly got lost. Is the following scenario true for which (if any) browsers, and is there definitive documentation for either Firefox or IE that states as much?: First Time Around: 1.1: User enters link to site A, or clicks on a link directed at Site A 1.2: Browser interprets link at Site A, first time, no cache. Sends GET to Site A. 1.2: Site A responds with 301 Redirect to Site B 1.3: Browser sends GET to Site B. Any Subsequent Times Around: 2.2: User clicks on a link directed at Site A 2.2: Browser sees that, due to a past 301 redirect, Site A should now be Site B. 2.3: Without initiating any request whatsoever at Site A, browser initiates GET at Site B.

    Read the article

  • Good language to learn in order to build small websites

    - by mkoryak
    I want to start building websites and charging people for them! My problem is that the stack that know well does not lend itself to quick development, or cheap hosting. I am looking for languages that satisfy the following criteria: Fast to develop in Can find cheap hosting for it Bonus points if it can also be 'enterprisey'

    Read the article

  • URL shortening: using inode as short name?

    - by Licky Lindsay
    The site I am working on wants to generate its own shortened URLs rather than rely on a third party like tinyurl or bit.ly. Obviously I could keep a running count new URLs as they are added to the site and use that to generate the short URLs. But I am trying to avoid that if possible since it seems like a lot of work just to make this one thing work. As the things that need short URLs are all real physical files on the webserver my current solution is to use their inode numbers as those are already generated for me ready to use and guaranteed to be unique. function short_name($file) { $ino = @fileinode($file); $s = base_convert($ino, 10, 36); return $s; } This seems to work. Question is, what can I do to make the short URL even shorter? On the system where this is being used, the inodes for newly added files are in a range that makes the function above return a string 7 characters long. Can I safely throw away some (half?) of the bits of the inode? And if so, should it be the high bits or the low bits? I thought of using the crc32 of the filename, but that actually makes my short names longer than using the inode. Would something like this have any risk of collisions? I've been able to get down to single digits by picking the right value of "$referencefile". function short_name($file) { $ino = @fileinode($file); // arbitrarily selected pre-existing file, // as all newer files will have higher inodes $ino = $ino - @fileinode($referencefile); $s = base_convert($ino, 10, 36); return $s; }

    Read the article

  • Render web page to picture.

    - by Steck
    I have the string with URL. For example "http://google.com". Is there are any way to download and render this page to picture file? ("test.jpg") I tried to use WebBrowser control, to download and render picture, but it only works when WebBrowser placed in displayed form. In other ways it's render only black rectangle. But i want to render picture without any visual effect (creating, activating form etc.)

    Read the article

  • Error message when running a Sinatra application on Windows Vista.

    - by Adam Siddhi
    I was following a video tutorial by Adam Keys about how to make a URL shortener app in Sinatra. The code that is giving me problems is located here http://pastie.org/958644 So when I ran it I got this error: shortener.rb:12: syntax error, unexpected $undefined, expecting $end @@ home I typed it exactly as Adam said but this keeps happening. What could the issue be? By the way, I am running ruby 1.9.1p243 (2009-07-16 revision 24175) [i386-mingw32] on Windows Vista. Thanks, Adam

    Read the article

  • CSS box around box technique

    - by webzide
    Dear Experts, I was trying to make an CSS division box with content in it as well as a border around it. Instead of using the box-border technique, I was trying out a new box to box technique instead. <html> <head> <style type="text/css"> #outer{ height: 20px; width: 20px; background-color:#233D78; } #inner{ height:18px; width: 18px; background-color: #FFF; font-size: 1em; text-align:center; font-family:'Bookman Old Style', serif; padding: 0px; margin-top: 1px; margin-right:auto; margin-left:auto; margin-bottom:1px; vertical-align:middle; } </style> </head> <body> <div id="outer"><div id="inner">TEXT</div></div> </body> </html> Somehow the borders are just not showing up right with FireFox. I tried everything. Setting up the Paddings of both boxes, margin, and messing around with the width. TO be honest, it took me around 30min to do this and I still can't get it right :( I know that a way to achieve the same result would be setting up a border around just one box. But I just wanna learn this box around box background-color technique. THanks in advance

    Read the article

  • Using Entity Framework as Data Access Layer

    - by AJM
    In my ASP.NET application I want to implement by data acess layer using the Entity framweok so I can use it as an ORM tool. But I dont want the rest of the application to care that I'm using this or be polluted by anything entity frameowrk specific. I cant seem to find anyone who is using the entity framework exclusively in their Data access layer so I'm keen to see any online examples of this/ experience people have.

    Read the article

  • need to open an image open in web browser

    - by manish
    byte.eml file is having image base64 encoded value ..and i am tring to open it in browser ...but this is not populating image file....plz help me out.. this is code... Dim oFile As System.IO.File Dim orEAD As System.IO.StreamReader orEAD = oFile.OpenText("E:\mailbox\P3_hemantd.mbx\byte.eml") Dim content As String content = "" ''Dim intsinglechr As Integer ''Dim csinglechr As String While orEAD.Peek <> -1 content = content & Chr(orEAD.Read) content = Replace(content, vbCrLf, "") content = Replace(content, vbTab, "") content = Replace(content, " ", "") End While Response.ContentType = "image/jpeg" Response.BinaryWrite(Convert.FromBase64String(content))

    Read the article

  • search engine crawling frequency

    - by Aditya Pratap Singh
    I want to design a search engine for news websites ie. download various article pages from these websites, index the pages, and answer search queries on the index. I want a short pseudocode to find an appropriate crawling frequency -- i do not want to crawl too often because the website may not have changed, and do not want to crawl too infrequently because index would then be out of date. Assume that crawling code looks as follows while(1) { sleep(sleep_interval); // sleep for sleep_interval crawl(website); // crawls the entire website diff = diff(currently_crawled_website, previously_crawled_website); // returns a % value of difference between the latest and previous crawls of the website sleep_interval = infer_sleep_interval(diff, sleep_interval); } looking for a pseudocode for the infer_sleep_interval method: long sleep_interval infer_sleep_interval(int diff_percentage,long previous_sleep_interval) { ... ... ... } i want to design method which adaptively alters the sleeping interval based on the update frequency of the website.

    Read the article

  • Using SalesForce's Web Service to create and set the type of a Task

    - by Alan Williamson
    I am successfully creating a Task using the SalesForce API SOAP API through Java. However, my problem is that I can't seem to set the Type of it. They all default to "Call" but I really want them to be "Email". Can someone point me in the direction of where I can do this? I think it is to do with RecordTypeMapping, but i am somewhat confused as to how to use this in my Java code to look up the particular one for Task type. I feel I have got so close with this. I have the correct WSDL that is giving me the extra method on the Task.java class, but no matter what I pass in, it dies. This doesn't seem to be a huge ask, yet i am perplexed as to which dots to join to get it to work Any help would be appreciated. thanks

    Read the article

  • Android WebView seems to ignore "viewport" information on web pages

    - by Evan
    I have a website that is using the viewport META tag to tell mobile browsers how to display content ( ). Viewing the page in the Android browser looks correct (and iPhone, etc). When I load the page into a WebView component in an android Application, the WebView ignores the "VIEWPORT" tag, and renders the page at "full" resolution, which is zoomed-in in this case.

    Read the article

  • ffmpeg web server settings

    - by uzay95
    I can easly upload video file to the server with SaveAs method. But when i wanted to convert video file to .flv file i can't see any output file in the folder that i expected to find. Actually i did it in my VS.NET 2008 IDE but after i published the files to the server, when i wrote the url of that internet site i couldn't see the files(in the remote machine i can see exported file and its converted file by calling http://localhost/FileUpload.aspx )

    Read the article

< Previous Page | 588 589 590 591 592 593 594 595 596 597 598 599  | Next Page >