Search Results

Search found 29495 results on 1180 pages for 'cross site scripting'.

Page 843/1180 | < Previous Page | 839 840 841 842 843 844 845 846 847 848 849 850  | Next Page >

  • Windows 2008 Enterprise License can't activate Standard Edition

    - by starchx
    We have downloaded and installed the Windows 2008 Standard R2 edition months ago. Now the server is in production. We signed up the Microsoft Partnership Action Pack subscription last week and get a license for Windows 2008 Enterprise edtion. I am trying to activate the standard edition with Enterprise key as advsied here:http://serverfault.com/questions/318968/upgrade-domain-controller-sku-from-server-2008-r2-standard-to-enterprise , but failed. Is it because windows 2008 we have is different? (downloaded from MS site with eval license). Thanks. tim

    Read the article

  • How can I use the proxy settings on Epic privacy browser to log on to Facebook?

    - by EddieN120
    I love the Epic privacy browser because it is built from the ground up to enhance privacy. It's built on Chromium but because it has stripped out all code that tracks users across the Internet, pages load faster and things work snappier. With one click you can enable a proxy to hide your IP address, sort of like Chrome's "Incognito" mode on steroids. But there's a problem: if I load Epic, go to facebook.com, log in, and then click the proxy button, I can use Facebook for a while. But eventually, Facebook would throw up an error screen, saying that it thinks that my account has been hacked, and then it would make me verify my identity, force me to change my password, etc. I've had to change my password four times in as many days, which is very annoying. Now I turn on the proxy for browsing on to every other site but Facebook. Question: how can I use the proxy settings on Epic privacy browser to successfully log onto and use Facebook?

    Read the article

  • "Turn Off The Lights" on any website?

    - by gojira
    On Youtube, there is this nice button (easy to overlook - top left of the video) which lets one "turn off the lights": the site background changes from white to black, the text color changes from black to grey. There is an unrelated plug-in for Firefox called "Turned Off The Lights", which has a very similar functionality. This makes websites so much easier to read. However, both technologies only work on YouTube. Is there anything to achive the same effect for all websites? Preferably with Firefox? I.e.: I want to have very dark background and light text color on all websites viewed with Firefox, how can I do that?

    Read the article

  • Sniff packets using tcpdump

    - by denisk
    I have a completely noob question. I want to see all packets that come to my computer from particular site (google.com). So I start tcpdump sudo tcpdump -i eth0 host google.com and enter google.com in a browser and hit enter - nothing gets captured. I can't figure out why it happen. What do I do wrong? Edit It appeared that I was listening to the wrong interface. I had changed eth0 to any and it worked. It was ppp1 that needed listening. Thanks for your answers!

    Read the article

  • Obfuscation is not a panacea

    - by simonc
    So, you want to obfuscate your .NET application. My question to you is: Why? What are your aims when your obfuscate your application? To protect your IP & algorithms? Prevent crackers from breaking your licensing? Your boss says you need to? To give you a warm fuzzy feeling inside? Obfuscating code correctly can be tricky, it can break your app if applied incorrectly, it can cause problems down the line. Let me be clear - there are some very good reasons why you would want to obfuscate your .NET application. However, you shouldn't be obfuscating for the sake of obfuscating. Security through Obfuscation? Once your application has been installed on a user’s computer, you no longer control it. If they do not want to pay for your application, then nothing can stop them from cracking it, even if the time cost to them is much greater than the cost of actually paying for it. Some people will not pay for software, even if it takes them a month to crack a $30 app. And once it is cracked, there is nothing stopping them from putting the result up on the internet. There should be nothing suprising about this; there is no software protection available for general-purpose computers that cannot be cracked by a sufficiently determined attacker. Only by completely controlling the entire stack – software, hardware, and the internet connection, can you have even a chance to be uncrackable. And even then, someone somewhere will still have a go, and probably succeed. Even high-end cryptoprocessors have known vulnerabilities that can be exploited by someone with a scanning electron microscope and lots of free time. So, then, why use obfuscation? Well, the primary reason is to protect your IP. What obfuscation is very good at is hiding the overall structure of your program, so that it’s very hard to figure out what exactly the code is doing at any one time, what context it is running in, and how it fits in with the rest of the application; all of which you need to do to understand how the application operates. This is completely different to cracking an application, where you simply have to find a single toggle that determines whether the application is licensed or not, and flip it without the rest of the application noticing. However, again, there are limitations. An obfuscated application still has to run in the same way, and do the same thing, as the original unobfuscated application. This means that some of the protections applied to the obfuscated assembly have to be undone at runtime, else it would not run on the CLR and do the same thing. And, again, since we don’t control the environment the application is run on, there is nothing stopping a user from undoing those protections manually, and reversing some of the obfuscation. It’s a perpetual arms race, and it always will be. We have plenty of ideas lined about new protections, and the new protections added in SA 6.6 (method parent obfuscation and a new control flow obfuscation level) are specifically designed to be harder to reverse and reconstruct the original structure. So then, by all means, obfuscate your application if you want to protect the algorithms and what the application does. That’s what SmartAssembly is designed to do. But make sure you are clear what a .NET obfuscator can and cannot protect you against, and don’t expect your obfuscated application to be uncrackable. Someone, somewhere, will crack your application if they want to and they don’t have anything better to do with their time. The best we can do is dissuade the casual crackers and make it much more difficult for the serious ones. Cross posted from Simple Talk.

    Read the article

  • Elevate the weight of browsing history in Google Chrome's autocomplete

    - by maayank
    Google Chrome has the feature of auto-completing web addresses while you type them in the address bar. Alas, it gives absurdly more weight to Google's own auto-suggest v.s. my own browsing history, which seems a bit foolish - if I regularly (i.e. twice a week) check a certain website with the keywords "foo bar ponies" in its url, it is reasonable to expect that I will want to visit that site again and not other sites. While a bit subjective, to the very least I would expect such URLs to be in the list Chrome suggests, even if not at the top. Is there some plugin/secret option that alters the default behavior?

    Read the article

  • Mcafee Auto-update from UNC path problem

    - by Vicky
    I have a network with 50 computers with no internet access. So instead of updating in each of them using dat file individually I tried to create a shared folder in server, and created a UNC in site repository. I downloaded the file DAT Package For Use with Mcafee AutoUpdate Architect & ePO 3.0 from http://www.mcafee.com/apps/downloads/security-updates/security-updates.aspx. When I try to update it is giving an error Error occurred while downloading file SiteStat.xml. So how fix it?

    Read the article

  • Webcam on Sony Vaio on Win 7 problem

    - by Norm
    Anyone know where I can get a Win 7 32Bit driver for my Web Cam (Vaio VGN-CR11H/B) ? I changed operating system from Chinese Vista to English Win 7 Update: Windows 7 will not detect the cam, but I just read the Sony Europe site and it gives me some ideas to try. I can dual boot this Sony laptop: with Chinese Vista, the camera works but IE8 does not work. I cannot read Chinese to fix the Internet access issue, but I can get the computer online and Skype works. When I boot with the new Win 7, IE 8 works great, but the cam doesn't work (mic and speakers work). Very perplexing problem.

    Read the article

  • Why is it good to have website content files on a separate drive other than system (OS) drive?

    - by Jeffrey
    I am wondering what benefits will give me to move all website content files from the default inetpub directory (C:) to something like D:\wwwroot. By default IIS creates separate application pool for each website and I am using the built-in user and group (IURS) as the authentication method. I’ve made sure each site directory has the appropriate permission settings so I am not sure what benefits I will gain. Some of the environment settings are as below: VMWare Windows 2008 R2 64 IIS 7.5 C:\inetpub\site1 C:\inetpub\site2 Also as this article (moving the iis7 inetpub directory to a different drive) points out, not sure if it's worth the trouble to migrate files to a different drive: PLEASE BE AWARE OF THE FOLLOWING: WINDOWS SERVICING EVENTS (I.E. HOTFIXES AND SERVICE PACKS) WOULD STILL REPLACE FILES IN THE ORIGINAL DIRECTORIES. THE LIKELIHOOD THAT FILES IN THE INETPUB DIRECTORIES HAVE TO BE REPLACED BY SERVICING IS LOW BUT FOR THIS REASON DELETING THE ORIGINAL DIRECTORIES IS NOT POSSIBLE.

    Read the article

  • Why my webservice not live on internet? [closed]

    - by blankon91
    I've windows server that goes live on internet, (e.g. www.mysite.com). Then I want to create another site with different port (e.g. www.mysite.com:502). I've create that and it works when I access it on local network, but when I access it from outside of local network (internet) the www.mysite.com:502 can't accessed but the www.mysite.com can accessed. what should I do to make www.mysite.com:502 goes online? I use windows server 2008 standard

    Read the article

  • SEO and internal links

    - by hanazair
    I'm fairly new to SEO and although I've read many articles on the topic I still don't have a clear idea of how to get my client's website get to the first page of Google Search. I run MOZ competitor analysis and see that a competitor that comes up at the top of Google Search has approximately same Domain Authority, Domain Moz Rank and Trust. They have 8 External Linking Root Domains while my client's site has five. Yet the competitor comes up as one of the top sites on the first page, and my client's side is on page #3. Then I noticed one drastic difference in competitor's ranking and that is Total Links. He has 1,388! I don't understand how this could be a positive factor in Search Engine ranking and how can they legitimately have 1,388 links (while only 14 of those are external). Another competitor who is #2 in search engine rankings has 773 links total with only 14 external links. It seems fishy, but yet there they are - at the top of the search engine results. Is that some current way to trick Search Engines? What to do if I'd like to get my client's website onto the first page by some legitimate means? Thanks.

    Read the article

  • Mail being delivered to both old and new server.

    - by Duane
    Two days ago I migrated our old site to a new server. I turned down the ttl before I did it and some email is still going to our old server and some to our new after a little over 48 hours. Both these are on shared hostgator sites. From what I can tell everything is set up properly. An nslookup points to the right ip address. Any insight? Support just tells me that everything is set up properly and I have to wait for propagation. I've never had it take over 48 hours though.

    Read the article

  • 25 years old and considering a career change...possible? practical?

    - by mq330
    Hi all, I'm new to this site and new to programming as well. I've spent some time going through an intro cs book that uses python as the language of choice. I find the exercises interesting and engaging and I generally have had a favorable experience programming so far. I've gone through some of the basics with python like writing simple programs, basics of GUIs, manipulating strings, lists, defining functions, etc. And I've always loved technology. Although I've never done any real hardcore programming yet, I was inclined to building websites from a very young age but I never really developed my skills. Now, the thing is I'm 25, I have my bacholors in environmental studies and two masters degrees in urban planning and landscape architecture respectively. I know, it would be quite a departure to pursue a career in programming at this point. Currently, I'm working as a geographic information systems intern. I've taken some GIS classes and have a lot of experience with making maps, doing spatial analysis etc. So what I'm thinking is maybe I can learn some solid programming skills and apply these skills in the field of GIS. From what I've seen, .net languages are the norm in this arena. Could you perhaps provide some guidance to me in terms of what languages I should focus on or courses I should take at this point? What about for building web mapping applications? Also, I was thinking about getting a certificate in programming from a university extension program. Do you think it would be worth it? And furthermore, do you think potential employers would be interested in hiring someone like me (once I get a couple of languages down pretty well) as an intern or in an entry level position? I'll be living in the bay area so I feel that there should be decent opportunities even though I don't have a b.s. in cs.

    Read the article

  • Creating Engaging Online Experiences is Easy and Intuitive for Marketers with Oracle WebCenter Sites 11g

    - by Christie Flanagan
    Last month, we announced the availability of Oracle WebCenter Sites 11g, the latest release of our web experience management solution. This new release is really geared toward enabling marketers and business users to drive customer acquisition and brand loyalty by simplifying the whole process of creating, managing and optimizing engaging online experiences.  To show you just how this works, we’ve created the video below which takes you through the tasks a typical marketer might execute using Oracle WebCenter Sites to manage their online presence -- everything from page editing to page creation, right on through to optimizing the mobile experience and moderating user-generated comments and reviews is covered here. I hope this video has give you a flavor for just how easy and intuitive it is for marketers and other business users to manage engaging and interactive online experiences using Oracle WebCenter Sites.  To see more about the new release, please check out the recording of our launch webcast. On Demand Webcast - Introducing Oracle WebCenter Sites: Transforming the Online Experience Enabling marketers and business users is a key requirement for creating and managing contextually relevant, social, and interactive online experiences. Oracle WebCenter Sites transforms the online experience into one that is simple and intuitive to manage as a content contributor, encourages interaction between site visitors and their social networks, and provides marketers with automated targeting options for optimizing online engagement. View this webcast now to learn more.

    Read the article

  • web spidering/crawling, can i do it or just search engines?

    - by bboyreason
    i already had a question answered about web-scraping with wget. but as i read a little more, i realize i may be looking for a web-crawling program. particularly the part about web-crawlers being able to get specific data like links or, in my case, products. all of the products on my site have the following naming convention, website.com/uniqueAlphaNumericID.html as far as i know, no dynamic content generation is being used and only one page per one item in the above format. should i just be thinking about: wget website.com | grep *.html or should i be looking into spiders/crawlers?

    Read the article

  • Standalone server setup for compute capacity

    - by mikera
    I'm developing an application for my company that will require a lot of compute capacity (running some very big mathematical calculations), and looking for some form of server setup to do this. For various reasons, we want to run this on-site in our office rather than hosting it externally. It's been a while since I last had to set up my own servers so I thought I would tap into the collective wisdom of serverfault! My broad requirements are: Budget $30-50k, with an aim to get as much compute capacity as possible for that budget 64-bit servers suitable to run Ubuntu Linux + Java Some relatively standalone rack that can be installed in secure office space Fast/low latency network connections between the servers, but don't really care about connectivity to the outside world Storage capacity shared between the servers - they don't necessarily need their own storage providing they can be booted from a common image Downtime can be tolerated (since the calculations are run in batch mode) The software itself is fault-tolerant, so there is no need for extra resiliency in the server setup (cheap replaceable commodity parts will be fine in general) Given these requirements what kind of setup would you recommend and why?

    Read the article

  • Is there a way to save MS Word document as HTML w/o the ms proprietary stuff?

    - by sequoia mcdowell
    So normally I wouldn't use this feature ("Save as Web Page") but I have large documents from clients they just want put on their site as HTML, and formatting it all by hand seems like a waste of time. I have tried "save as webpage" in Word 2007, but it produces all sorts of bad stuff. To wit: <b style='mso-bidi-font-weight:normal'> <span style="mso-spacerun: yes"> as well as a large block of XML formatting info: <!--[if gte mso 9]><xml> <o:DocumentProperties> <o:Subject> </o:Subject> <o:Author> </o:Author> <o:Keywords> </o:Keywords> ... As I said, formatting it all by hand seems like a waste of time, but the way MS exports currently just has too much cruft. Is there a way to export MS Word doc as html without all this?

    Read the article

  • block access to certain website types

    - by frustrated teacher
    Need to block access to certain website types without listing each URL to block. Students at secondary school are going to porn sites. Need to be able to block all such access without having to list each possible site URL. Having the Content -- Ratings tab set to None for all categories on the ratings files listed on my computers does not prevent access. Unchecking users may access sites with no rating, even with the security settings set to High, still allows the porn sites to come up. If that is checked, then ONLY listed sites can open and students would not be able to do any research via google, for example. I would rather not have to continue checking each computer and blocking sites as they find them.

    Read the article

  • Ubuntu 11.10 w/ Cinnamon - Icons are gone and pictures do not display

    - by B U N N Y
    I'm new on this site and new to using Ubuntu and all. I currently have Ubuntu 11.10 and Cinnamon along with it. I've recently downloaded only a few applications, like the Adobe shockwave as well as a few games. After the downloads, the icons for the windows (the close, minimize, and maximize) are nothing but streaks of colours, and now just plain black, picture previews in my picture folder is gone, some icons on the panels are missing, and icons in the control center are gone. I'm sure more are gone, but that's the ones that I see often. I'm not sure what caused this to happen?? Please get back to me soon! I'm worried about what has happened. ): This is my only laptop to get online. I have no other way besides the library but that takes too long.. EDIT: Okay, I had asked my sister for some help and she helped me restore my minimize, maximize, and close button but I still haven't figured out how to show the icons in the cinnamon dock (which I have set at the top of the window) and how to have my pictures show up again. All of my pictures are just blank and I can't even preview them. They still have their file names but that is it. I'm not sure what has happened. I haven't downloaded any new apps as of the new edit.

    Read the article

  • What does my blog says about me?

    - by subodhnpushpak
    Somebody said that “a person is as good as he/she THINKS”. Recently, over conversation, one of my close “blog friend” remarked that a blog post revealed so much about the person who has written it; and though we never met physically / or even talked over phone… (we have only exchanged technical talks over chats and blogs); this friend of mine correctly predicted my personality and even my habits!! I also came across with this site http://www.typealyzer.com/?lang=en. Just put a blog address and it will predict the bloggers mindset!! Here is what I get when I put across my blog ..   very flattering… I must say. Why don’t you discover yours?? Do share the results with me.. BTW.. now a days I am working on Windows Phone 7 and am loving it. Have lots of gotchas to blog but simply no time (only 14 hrs a day )

    Read the article

  • Illustrator "Save for Web & Devices" returning crappy, pixelated images

    - by Tory Waterman
    I'm trying to create a nice title for my webpage... a big white title to sit on a black background. I'm using Illustrator to do so. When I create it, it looks nice, but when I hit "save for web & devices", it comes out looking like a pixelated piece of crap on the site. Is there some setting I need to change to make Illustrator save a higher resolution image? Thanks EDIT I understand, from looking at some other posts, that this may be a result of "posterization" or "dither", but this is only a plain white image so I don't how this results in a colors problem. (I could be completely misinterpreting these terms) EDIT Figured version might be important... I'm using CS5.1

    Read the article

  • public_html permissions for local development

    - by maGz
    I know this question has popped up a couple times, but I can't seem to find a definitive answer to my issue, so please bear with me. I have Ubuntu Server 12.04 setup in VirtualBox for PHP development and testing (Drupal plus other PHP sites using Yii framework). My question is in 3 parts... 1) If I create a public_html folder under /home/myuser, do I need to give ownership of that folder to the Apache www-data group? If so, are there any specific permissions I should be setting? 755? (Btw, I am following this guide to create the public_html directory and set up multiple virtual hosts per site I create and test) I previously had all of my sites under /var/www, but ran into massive permission denied errors whenever I tried to sFTP to it, either through FileZilla or PhpStorm. This is what I had previously done: sudo chgrp www-data /var/www sudo chmod -R 775 /var/www sudo chmod -R g+s /var/www sudo usermod -G www-data [my_ftp_user] 2) The second part of my question is this: If I create my PHP project and files in Windows through PhpStorm, and then upload via sFTP, will permissions get affected? 3) Once I am satisfied with my developed project, would it be advisable to move and test them under /var/www to see how it would fair in a production-ish environment? I would really appreciate the help and advice here. I'm learning more as I go along, but dealing with Linux files and permissions is a bit of a new ballgame for me! Thank you

    Read the article

  • External TV only allows 640x480 using VGA cable

    - by EdC
    I have Ubuntu precise installed on a MacBook Pro. When connecting to an external television using a VGA style cable the highest resolution available seems to be 640x480. I'm using the proprietary NVIDEA Accelerated Graphics Driver (version current) and attempting to setup the TV using the nvidea-settings tool. I know all the equipment supports higher resolutions because of the following: Prior to rebuilding my machine with precise, the same TV, cabel & adapter gave higher resolutions The same machine & TV when used with a HDMI cable allows higher resolutions Windows laptops with the same cable & TV are able to achieve higher resolutions I'm not sure what to look for next, there are some similar FAQ questions but I'm not sufficiently familiar with the areas involved to know if these are my issue or are safe to try. Specifically: colord problem & my monitor resolution 640*480 with no other options! - is the closest, but as I'm having no problems with the main laptop monitor I'm not sure the described symptoms apply. I can't work out from the answer or Bumblebee site how to determine if nvidia optimus applies to my laptop. Why isn't my monitor's native resolution appearing as an option? - suggests configuring X by hand, but doesn't entirely match the symptoms I'm seeing

    Read the article

  • How do I remove the tool tips on the launch bar?

    - by Sephethus
    The title is the question. These tool tips stay until I try to click past them. They're annoying since they constantly pop up and block the view of what I'm trying to do. Unfortunately I need the launch bar because Ubuntu is running on VMware and the console does not allow me to use the keyboard for switching tasks (to my knowledge). How do I disable them? I'd post an image as an example, but this site will not let me. UPDATE: unity 5.16.0 UPDATE 2: I discovered that this may only be a problem with users who run Ubuntu on a full screen VMware console that is situated on the right monitor. When the mouse is moved to the left monitor the tooltips popup and remain until the mouse is clicked twice in the VMware console window to make it active. Unfortunately my problem is one involving a rare situation I think. However, I'd love to be able to disable these tool tips if possible. It would also be nice if new features were added that can allow further customization of the launch bar.

    Read the article

  • Client-side certificates

    - by walshms
    My company purchased a wildcard certificate from a vendor. This certificate was successfully configured with Apache 2.2 to secure a subdomain. Everything on the SSL side works. Now I'm required to generate x509 client-side certificates to issue for this subdomain. I'm following along this page: (http://www.vanemery.com/Linux/Apache/apache-SSL.html), starting with "Creating Client Certificates for Authentication". I've generated the p12 files and successfully imported them into Firefox. When I browse to the site now, I get an error in FireFox that says "The connection to the server was reset while the page was loading." I think my problem is coming from not signing the client-side correctly. When I sign the client-side certificate, I'm using the PEM file (RapidSSL_CA_bundle.pem) from RapidSSL (who we bought the certificate from) for the -CA argument. For the -CAkey argument, I'm using the private key of the server. Is this correct?

    Read the article

< Previous Page | 839 840 841 842 843 844 845 846 847 848 849 850  | Next Page >