Search Results

Search found 10447 results on 418 pages for 'blog blomqvist no'.

Page 260/418 | < Previous Page | 256 257 258 259 260 261 262 263 264 265 266 267  | Next Page >

  • NGinX config for Django and Wordpress in subdirectory

    - by Helmut
    I need to set up a Django site at the root of a domain, but then have a Wordpress installation in a subdirectly (e.g. /blog/). How would one configure NGinX to do this? "Pretty" URLs have to work for Wordpress as well. For Django I am using Gunicorn, which is already configured. From NGinX I would call "proxy_pass" to direct to that. PHP is run via FPM. Considering the restrictions above, how would I configure NGinX? Any help would be appreciated! Thanks.

    Read the article

  • FeedValidator & Feedburner get 404 when accessing wordpress RSS feeds when permalinks are enabled.

    - by Wazbaur
    I'm helping a friend set up a self-hosted Wordpress blog + feedburner and I'm seeing a problem with the feeds that I'm finding somewhat mysterious. Using the default permalink structure (e.g., ?p=123) everything works as expected; I can follow the feed in Google reader, navigate to it manually, and set it up in feedburner. However, once I switch away from the default permalink structure, feedburner and feedvalidator both report that accessing the feed is returning HTTP-404 and Google reader no longer shows new posts (I'm assuming for the same reason), but I can navigate to the feed using a browser. When I do that it appears as though nothing is wrong; there is a feed there and it contains all the posts I expect it to have. I've re-started the feedburner & reader set-up from the beginning after changing the link structure, so I don't think they're doing anything silly like looking at the feed at its old address. I've seen people with similar problems in various other places but there doesn't seem to be a good answer anywhere.

    Read the article

  • Auto-scaling EC2 Servers and Updating Code

    - by jstats
    We've come to the point where we need to set up autoscaling for our web server and I'm unsure how to go about the process of scaling servers and updating the the existing code without remaking a new AMI and changing the autoscale config to use it. I've read a bit about people bundling the new code and uploading it to s3 and having new servers grab the bundle on boot up but that doesn't seem all that pleasant either. Currently the web app's files live in a git repo, and when we update the code, we push it to github, ssh into the web app and run a hook to bring down the latest code. So I was thinking that another option could be to just run that hook on an hourly or daily cron task. Unfortunately that doesn't cover everything type of update (for example new blog posts' images and such which aren't included in the git repo) but it's something. Could anyone provide some advice on what a common solution is or anything as to why my proposed solution is a bad idea? Thanks all

    Read the article

  • Why can't SVN checkout into a virtualbox shared folder?

    - by Alex Waters
    I am trying to checkout into the virtualbox shared folder with svn 1.7 in ubuntu 12.04 running as a guest on a windows 7 host. I had read that this error was a 1.6 problem, and updated - but am still receiving the error: svn: E000071: Can't move '/mnt/hostShare/code/www/.svn/tmp/svn-hsOG5X' to '/mnt/hostShare/code/www/trunk/statement.aspx?d=201108': Protocol error I found this blog post about the same error in a mac environment, but am finding that changing the folder/file permissions does nothing. vim .svn/entires just has the number 12 - does this need to be changed? Thank you for any assistance! (just another reason for why I prefer git...)

    Read the article

  • linking Office 365 and ADFS on Azure

    - by Gaurav
    i am trying to link my ADFS to Office 365 in order to set up Single Sign On for Office 365. I am going over this blog post. at Step 5, i am supposed to run Powershell commands so i can connect the ADFS to Office 365. Set-MsolAdfscontext -Computer <AD FS server FQDN> Convert-MsolDomainToFederated -DomainName <domain name> i am unable to determine the FQDN for the ADFS server. i searched around a lot on the web, but was not able to locate the solution. i am just getting started with all configuration of ADFS and Office 365, but was not able to find out a solution. any answers?

    Read the article

  • Playing/extracting audio file from PDF

    - by ravl1084
    I use Ubuntu and I have a PDF file that contains an audio annotation. It won't play on Okular, it treats it as a text annotation. Following an old blog post where the poster created a small C script to extract the audio didn't work either, I suspect the format of these audio annotations has changed. Using the information on it I managed to uncompress the PDF and with vim, I found the audio data in the file. I tried copying this into its own file and changed the extension from mp3, wav, mid, but none of them would play. Is there a way of achieving this?

    Read the article

  • Why are PNG-8 files mangled when opened in Photoshop?

    - by Daniel Beardsley
    Why are some 32 bit PNGs opened in Photoshop with Indexed Colors and no transparency? For instance, I grabbed a png icon file of the Stack Overflow logo at: http://blog.stackoverflow.com/wp-content/uploads/icon-so.png When opening it in Photoshop CS3, it apparently treats it as indexed color and gets rid of the alpha channel. The image on the right is a screen grab of the icon. Changing the Image mode in Photoshop to RGB doesn't change the image at all. I've tried this with a few other PNGs and it seems hit or miss. When viewed in other programs, it displays fine. left:png opened in Photoshop, right:screen grab of png from browser What gives?, does Photoshop not interpret the PNG file format correctly?

    Read the article

  • Access Linux Samba-Share from Snow Leopard

    - by Space_C0wb0y
    Hello there! I have a linux machine with a samba-share. Other windows machines can access that samba-share no problem. Now I have a new MacBook running Snow Leopard and I have no clue on how to access it. In finder it shows the linux-machine, but I have no idea how to access it. In some blog it said that this: `smbclient -U <username> -L <hostname>` can be used to list samba-shares. When I do it, I get `session setup failed: NT_STATUS_LOGON_FAILURE` I am at a loss here. Any help on how to approach that problem would be appreciated. Can you provide some hints or point me to some good howto?

    Read the article

  • Creating a seperate static content site for IIS7 and MVC

    - by JK01
    With reference to this serverfault blog post: A Few Speed Improvements where it talks about how static content for stackexchange is served from a separate cookieless domain... How would someone go about doing this on IIS7.5 for a ASP.NET MVC site? The plan so far: Register domain eg static.com, create a new website in IIS Manually copy the js / css / images folders from MVC as is so that they have the same paths on the new server Enable IIS gzip settings (js/css = high compression, images = none) Set caching with far future expiry dates <clientCache cacheControlCustom="public" /> in the web.config Never set any cookies on the static.com site Combine and minimize js / css Auto deploy changes in static content with WebDeploy Is this plan correct? And how can you use WebDeploy to deploy the whole web app to one server and then only the static items to another? I can see there is a similar question, but for apache: Creating a cookie-free domain to serve static content so it doesn't apply

    Read the article

  • git-receive-pack : command not found.

    - by Philippe Mongeau
    I made a git repo on a local machine with "git init --bare" and added it as the remote origin on the project on my main computer with ssh: git add remote origin [email protected]:repoName.git I was able to make a commit and push from my main computer to the other computer the day I created the repo, but today i tried and it didn't work. When I did "git push origin" it returned this error: bash: line 1: git-receive-pack: command not found fatal: The remote end hung up unexpectedly The two machines are mac the main one running Leopard and the server one running Tiger. I think it may be realted to the $PATH of git on the server but I'm not sure. i used theses instrution to create my git server: http://blog.commonthread.com/2008/4/14/setting-up-a-git-server

    Read the article

  • Different versions of 2.5" drives? Unscreawing Intel 320 for slimmer drive to fit newer thinkpads?

    - by hhh
    My x60s laptops come with about 1-2mm heigher 2.5" HDD than newer x220 models. This is totally stupid when one would like to reuse the old drive in the newer laptops. I bought newer 2.5" 320 Intel SSDs and they seem to have such 1-2mm gap to unscrew but I am unsure whether it is meant for opening. Could someone instruct what to do here? Look the manufacturer has started changing the old good 2.5" drives into slightly different versions, now it means slow compability issues to fix or totally new 2.5" drives. Ideas how to proceed? Unscrewing newer 320 drives or are some other versions of 2.5" drives meant for x220? Does there exist some sort of racks to get drives working between different comps? Perhaps booting from laplink is currently the easiest solution to get things working when changing harddrives between comps? Perhaps related http://www.zdnet.com/blog/hardware/seagate-announced-super-slim-25in-momentus-thin-hard-drive/6439

    Read the article

  • Building modules with linux kernel for custom flavor

    - by Chandan Apsangi
    I followed the instructions given in the link: http://blog.avirtualhome.com/how-to-compile-a-new-ubuntu-11-04-natty-kernel/ for building a custom kernel and booting it. Everything works fine, except that when building it, I used the option skipmodule=true (as given in this link), so I guess the modules are not built for this kernel. I'm saying this because after I installed the image (using dpkg -i), when I run 'ls -l' in the /lib/modules/2.6.38-11-i5 folder, all the files/folders except the 'kernel' folder have got updated (Btw, I had built it last time and it had worked fine,this is an update on the same kernel source). So is there some option/command that I missed that exports the modules as well? Thanks for your help.

    Read the article

  • Running 32 bit SQL Server 2005 on 64 bit Windows Server?

    - by TooFat
    If I have a 32 bit version of SQL Server 2005 running on a 64 bit Windows Server does the max amount of memory avail. the SQL Server process increase from 2gb to 4gb. In reading this blog entry by Mark Russinovich in which he states that "All Microsoft server products and data intensive executables in Windows are marked with the large address space awareness flag" and "Because the address space on 64-bit Windows is much larger than 4GB, something I’ll describe shortly, Windows can give 32-bit processes the maximum 4GB that they can address and use the rest for the operating system’s virtual memory." which leads me to believe that the answer is "yes" but I not totally confident.

    Read the article

  • MAMP and WordPress - Theme is gone

    - by ninjaboi21
    Hey SuperUsers, I hope this is the right section to post this kind of question. If it's not, I am sorry. But to my question: I created a working WordPress blog, running locally with MAMP Pro. I wanted to share my site, and thought I would go public with it, but self hosted. I signed up at DynDNS for my domain and got it working. People can connect, myself included of course. But here goes my problem: Everytime I try to access my website (**.dyndns.info:8888), I am able to access it, but the theme and pictures are gone, so it's only pure text with some links. I can provide a picture of the website if you want it? But what is it I am doing wrong? I tried to enable Web Sharing, but that didn't make a difference. Thank you so much if you can figure this out!

    Read the article

  • Is it safe to compress my Windows 7 %USERPROFILE%\AppData folder?

    - by Kev
    Having just read Scott Hanselman's latest blog entry, Guide to Freeing up Disk Space under Windows 7, he suggests turning on NTFS compression which I already do for a number of less travelled folders that contain static files such as downloads or images. However I am wondering if it's wise to turn on NTFS compression for the whole of my %USERPROFILE%\AppData folder? My system drive is a 128 GB SSD residing in a Dell Precision T5400 3Ghz Quad Core Xeon workstation so I ought not to notice the extra cycles used to compress and decompress files on their way to and from the disk. However would there be any good reasons not to do this? In fact could I safely compress the whole of my %USERPROFILE% folder?

    Read the article

  • How much CPU use is too much?

    - by Jonathan Sampson
    I've got a server that receives around a million unique visitors a month and I've recently began using Plesk to help monitor some of the vitals on the box itself. RAM I can make sense of, but I'm not really sure if my CPU usage is too high, low, or about average for this number of visitors. The server exists solely to serve up a somewhat hefty WordPress blog. This is one week. What types of things should I look out for? Some other information about this server follows: VCPU(s): 4, RAM: 6GB, HDD: 30GB, OS: Ubuntu Server 10.04 x86_64

    Read the article

  • Windows Server 2003 Router, Good approach [closed]

    - by jM2.me
    Possible Duplicate: Windows Server 2003 Router with PortForwarding Situation We have Verizon Fios 25/25 Internet Connection, Server acting as a router, and around 12 office computers. Task Portforward port 29000 from office computer. Problem Once I connected wan and lan cables I just had to set static lan ip (on server) and plug in switch with office computers into second nic. Then just right click on wan nic and select share internet connection. All office computers were assigned with an IP address 192.168.0.XXX and Gateway 192.168.0.1 (server). Now I have to open port and portfoward it from computer 192.168.0.190 (static ip, manual). Using this guide http://www.rosscode.com/blog/index.php?title=port_forwarding_in_windows_2003&more=1&c=1&tb=1&pb=1 I faced a problem. Before enabling "RRA" I had to unshare internet connection (wan interface) and was able to setup network. Now how do I setup a network within "RRA" and share internet with private network? Thank you much

    Read the article

  • IIS7 is gzipping files but not serving the gzipped version.

    - by ptrin
    By following a number of helpful blog posts I have configured IIS to gzip my static files. I have even enabled Failed Request Tracing and filtered to the 200 status code, and I can see the successful compression events taking place as well as the finished headers, which look like this: Headers="Content-Type: text/css Content-Encoding: gzip Last-Modified: Mon, 04 Oct 2010 17:35:08 GMT Accept-Ranges: bytes ETag: "02ef37cea63cb1:0" Vary: Accept-Encoding Server: Microsoft-IIS/7.5 X-Powered-By: ASP.NET " However, when I test in Fiddler and Firefox the Content-Encoding header is missing, and the file is not gzipped. This is a similar issue to this question which was never resolved. IIS is generating the gzipped files which I can see in C:\inetpub\temp\IIS Temporary Compressed Files . Does anyone know how I can troubleshoot this?

    Read the article

  • Problems with the backup

    - by marcodv
    I did a script which run around 4 o'clock in the morning, for backup all the mysql databases and the config file for 250 linux vm. The problem is that it tooks ages for complete and more than 50% of these vm, need more than 8 hours for complete. More or less all the vm had the same configuration,I mean Same amount of ram same amount of disk space same number of cpu Debian 6.0.5 I am saving these backup on amazon s3, because is the cheapest solutions that I've found. Now my questions is: Has anyone some solutions or suggestions about that? On one blog I've read that probably the ionice and nice combination could be good work around about that. any thought?

    Read the article

  • Have main website hosted on 3rd party while keeping Google Sites for Users

    - by vinnybozz
    Hi, I want a third party hosting my main site with PHP, MySQL, etc... I don't know which DNS records to modify. Is it possible to have the following mappings: www.example.com = 3rd party hosting blog.example.com = other 3rd party hosting mail.example.com = Google Mail docs.example.com = Google Docs sites.example.com = Google Sites sites.example.com/internal-site = Google Sites internal site ... Right now in TotalDNS, I have www = ghs.google.com. If I modify only this record to point to the IP provided by the 3rd party hosting, is it going to work ? Do I also need to add NameServers, remove the ones Google added ? Thx for the help

    Read the article

  • Is it safe to set MySQL isolation to "Read Uncommitted" (dirty reads) for typical Web usage? Even with replication?

    - by Continuation
    I'm working on a website with typical CRUD web usage pattern: similar to blogs or forums where users create/update contents and other users read the content. Seems like it's OK to set the database's isolation level to "Read Uncommitted" (dirty reads) in this case. My understanding of the general drawback of "Read Uncommitted" is that a reader may read uncommitted data that will later be rollbacked. In a CRUD blog/forum usage pattern, will there ever be any rollback? And even if there is, is there any major problem with reading uncommitted data? Right now I'm not using any replication, but in the future if I want to use replication (row-based, not statement-based) will a "Read Uncommitted" isolation level prevent me from doing so? What do you think? Has anyone tried using "Read Uncommitted" on their RDBMS?

    Read the article

  • How to auto-update a website mirror with exceptions to certain pages?

    - by tomatosalad
    I'm currently mirroring a website on my server. The site itself is rarely updated, but it is updated enough that info can become outdated quickly. I mirrored it first with wget, and this worked fine, but I made some changes: The original index.html used frames, but the site also provides a main.html which is essentially index.html but without frames. I deleted index.html and renamed main.html. I did not want to mirror the webchat, blog or forum, so I deleted those files and directories and made directories "blogs" "forum" and "chat" and placed a php redirect in each of those, redirecting visitors to the orignal site. I'd like to auto-update the mirror (maybe once every 24-72 hours), but preserve the changes I made. Is this possible? How would I go about doing it? I am completely clueless as to how. Thanks for any and all help! :)

    Read the article

  • How can I get Google to re-point its search entries to new domain?

    - by poolski
    My main .com domain registration lapsed and when I went to re-register it, I found that a domain reseller service squatted it and I've lost access to it. As I wasn't terribly keen on spending money on funding scammers and the like, I registered a .co.uk domain under the same name. Is there any way of getting Google to re-point all its indexed links to the new domain? It's been indexing my blog for a couple years now and while it's not too big a deal, I'd like to not have to start all over again. Also, searching for my site results in an old entry which is currently pointing at a "Apply for a Tax Break NOW!!!" page.

    Read the article

  • What would Stack Exchange's yearly expenses be if it were to be using a third party host?

    - by abel
    StackExchange manages it's own servers, as it should, but if SE were to be hosted on a 3rd party "cloud" hosting (like Amazon's), what would it's monthly / yearly expenses be(keeping everything else the same)? A detailed answer comparing it to the bills that Stackexchange boots currently (including power/property/staff) would help. (PS: I know that the blog is a good resource. I also understand that managing your own hosting is almost the same as setting up a hosting company and using it for your own needs. Plus is this a question for meta or does it fit within serverfault's purview?)

    Read the article

  • adding ftp user on ubuntu

    - by user46250
    I followed this tut http://www.trainsignal.com/blog/how-to-set-up-safe-ftp-in-linux to setup an ftp server with user account sudo mkdir -p /home/ftp/ftpuser sudo useradd ftpuser -d /home/ftp/ftpuser -s /bin/false sudo passwd ftpuser when I tried to connect with login ftpuser remotely it didn't work. It didn't work even with root UNLESS I removed root from ftpusers. I am confused ftpusers are the users NOT allowed to do ftp ? Where are the list of users allowed then and why can't I connect with ftpuser I created ?

    Read the article

< Previous Page | 256 257 258 259 260 261 262 263 264 265 266 267  | Next Page >