Search Results

Search found 19155 results on 767 pages for 'url redirection'.

Page 516/767 | < Previous Page | 512 513 514 515 516 517 518 519 520 521 522 523  | Next Page >

  • Plus sign ( + ) in a rewrite rule (.HTACCESS on Linux Server) (SOLVED)

    - by ienabellamy
    i've this snippet, but it doesn't works. Obviously is that plus sign + tha must be escape ? RewriteCond %{QUERY_STRING} ^orderby=position&controller=search&orderway=desc&search_query=t+max$ RewriteRule ^it/cerca$ http://www.wrs.sm/it/cerca?orderby=position&controller=search&orderway=desc&search_query=tmax? [R=301,L] i want to redirect this url http://www.wrs.sm/it/cerca?orderby=position&controller=search&orderway=desc&search_query=t+max to this http://www.wrs.sm/it/cerca?orderby=position&controller=search&orderway=desc&search_query=tmax Thanks ;) I found the solution, the correct rules are: RewriteCond %{QUERY_STRING} ^orderby=position&controller=search&orderway=desc&search_query=t\+max$ RewriteRule ^it/cerca$ http://www.wrs.sm/it/cerca?orderby=position&controller=search&orderway=desc&search_query=tmax [R=301,L]

    Read the article

  • XML Parsing Error at 1:1544. Error 4: not well-formed (invalid token)

    - by Steve
    I have installed Joomla 1.5.22 on a new hosting account, which doesn't have a domain yet, so it's public URL is http://cp-013.micron21.com/~annimac/ A message saying: XML Parsing Error at 1:1544. Error 4: not well-formed (invalid token). The source code for this message is: <dl id="system-message"> <dt class="error">Error</dt> <dd class="error message fade"> <ul> <li>XML Parsing Error at 1:1544. Error 4: not well-formed (invalid token)</li> </ul> </dd> </dl> There is nothing in /logs to indicate what the problem is. I have uploaded the following folders from a freshly unzipped copy of Joomla 1.5.22: administrator components includes language\en-GB libraries modules plugins templates\ja_purity xmlrpc and the issue remains. I have no custom or additional plugins, modules, or components installed. If I change templates, the problem remains. What is the problem?

    Read the article

  • Apache not serving pages stored in Subversion repository

    - by Stephen
    I've setup Apache and Subversion on an old PC, but Apache is not serving pages correctly, when I enter the address to my test site: http://HOME_IP_ADDRESS/test/index.html I just get a File Not Found error and the following output in the error log: File does not exist: /var/www/html/svn/repos/test but I know the file exists, when I enter the following URL into the browser: http://HOME_IP_ADDRESS/repos/test/index.html I just get a listing of the HTML. In my Apache config file I have the Document Root set as follows: DocumentRoot "/var/www/html/svn/repos" so I'm not sure what is going on, I have SVN installed and I think it may have something to do this. Edit * I changed the Document Root location, which helped as pages in the new location were served correctly, so the problem is with just serving the pages from the repository.

    Read the article

  • Junk host name while using nslookup, How to change it?

    - by Jayachandran
    Hi, My question is regarding the host name mapped to a machine's IP. I am new to Linux, I have a machine to test whether our j2ee based web application runs fine in linux box. we are able to deploy our application and successfully and even able to access the same in that machine. But when I tried accessing it through another machine in the same network, I am not able to access the URL using the hostname specified, instead I am able to access if I provide the ipaddress. I tried putting nslookup MyIP# command to know the name of the ip in the DNS server, where I got some junk value.com; when I tried to access with that name I am able to access the machine. Kindly provide me inputs to change the hostname of this machine in that domain server. Thanks in Advance!! Jay

    Read the article

  • phpMyAdmin not found on Windows

    - by Urban
    I installed wamp server 2.2 on Windows 7 x64. I clicked on its system tray icon and put the server online. Then i select phpmyadmin and I get a 404 page. Localhost successfully gives me the Apache 2 test page. Theres no way Im able to fix this. The php error log and the Apache error logs have no errors. Note: Ive tried all combinations of the url such as localhost/phpmyadmin/ and localhost/phpmyadmin and localhost/phpmyadmin/?

    Read the article

  • redirecting HTTPS requests to http in lighttpd

    - by chochim
    I have a lighttpd server running which has an SSL certificate installed. I would, due to certain reasons, like to forward all https: //www. requests to http: //www. My lighttpd code looks like as follows: $SERVER["socket"] == ":443" { ssl.engine = "enable" ssl.pemfile = "/path/to/pem/file" ssl.ca-file = "/path/to/ca/file" HTTP["host"] =~ "^www\.(.*)$" { url.redirect = ("^/(.*)" => "http://www.%1$1") } } Can you please point out the problem here. Another thing, what is the difference between %1 and $1 ?

    Read the article

  • Varnish with multiple hosts/subdomains

    - by jerhinesmith
    I'm new to Varnish, and I'm hoping it already does this "out of the box", but I'd like to clarify before I consider using it in production: Here's my setup: I have multiple sites running off of the same machine that vary by subdomain (i.e. user1.example.com, user2.example.com, etc.) Each "site" has a profile picture that has the same name (i.e. user1.example.com/profile.png, user2.example.com/profile.png) Will Varnish recognize these as separate resources and cache them accordingly? Or will I need to change something in the VCL to tell it include the full host url when looking up cache hits?

    Read the article

  • grub boot failed after upgrading to ubuntu 12.04 LTS

    - by user138021
    no such disk error occured. I tried to format and reinstall 12.04, the problem remained. I also repaired with boot-repair, the problem remained. http://paste.ubuntu.com/1224005/ is url of the details. server is dell R710, the bios is set to uefi and disk is raid0 gpt. I typed commands in grub rescue: ls (hd0, gpt2), `no such partition' is shown set, `prefix=((null),gpt2)/grub' is shown I don't know why /efi/ubuntu/grubx64.efi doesn't recognize disk. another strange thing is that there is only 1 file in /efi/ubuntu

    Read the article

  • Regarding the functioning of the HTTP server

    - by Luv
    As all of us uses the HTTP protocol for accessing the web, and request at the port number 80 of the server, now suppose i want to browse or download a file then i will request to the HTTP server by specifying the complete URL. If the server is having file, it would return it if it is not having the file then what will it do? I think it will make the request to the appropriate server which would be having this file and get the file and return to us. In this way does the HTTP server acts as an intermediate between the client and the server? Please correct me if i am wrong here. Many many thanx in advance

    Read the article

  • WGet a Page that Requires Logging in

    - by Synetech inc.
    I’m trying to figure out a way to use WGET or a similar tool so that I can schedule a web page to be downloaded regularly as a sort of updating log. The problem is that the page requires that I be logged in otherwise I get a different page, generic. Further, the page does not take login information as GET parameters in the URL, it uses POST to log in on the login page and cookies to save the login information that’s read by the regular page. I’m currently using GNU Wget 1.10.2 for Windows. I’ve tried using WGET’s cookie functionality but have had mixed results, usually skewing towards it not working. Can anyone please advise on a way to accomplish this? Thanks a lot.

    Read the article

  • Download web server structure with empty files

    - by golimar
    I want to make a mirror of a Web server, but downloading the actual files will take too long. So I thought of having just the directory and file structure, and when I need the actual contents of the file, I can download just that file. I have tried wget --spider URL and in a short time it has created in my local disk the directory structure with no files. But I've checked all of wget's or curl's switches and there is nothing like what I need. Can this be done with wget, curl or any other tool?

    Read the article

  • IE Search Provider: Specifying gTLD / Country-Specific Site

    - by jwa
    I am based in the UK, and as such typically use google.co.uk as my search engine. However, my employer is based in continental Europe, and thus my internet proxy is located overseas. As a result, IP geo-location presents a location outside of the UK. Google detects this, and as a result will redirect my searches from the address bar to a foreign Google domain. This leads to "local" answers having a higher ranking, many of which are not written in English language! Is there a specific search provider / URL I can give to IE which will use a specific gTLD of google (.co.uk), rather than performing the location-based redirect?

    Read the article

  • Rewrite 2 different directories with htaccess?

    - by jason
    I have a tricky problem (for me at least). I'm trying to rewrite / to a folder /webroot/www. I have some simple code and it works: RewriteRule ^$ /webroot/www/ [L] However at the same time if the URL starts with components, followed by anything else (ex. foo, as in /components/foo), and foo is an actual directory that exists inside components, I should rewrite to /components/foo/www instead. How can I achieve that? I can't seem to figure it out. I'm using Apache with .htaccess.

    Read the article

  • Cannot connect Solaris 10 to Windows 7

    - by user999353
    I'm trying to connect a Solaris VM (powered by VMware Player) to Windows Explorer in Windows 7. When I try to map the network drive I get the following: The specified server cannot perform the requested operation. I am using the following URL which has worked on a machine I used before. The only thing that has changed is the IP address of the Solaris machine. I am able to connect to the VM via putty. \\1.2.3.4\xxx\yyy I checked and I think samba is running.. root 375 1 0 09:53:39 ? 0:00 /usr/sfw/sbin/smbd -D root 376 375 0 09:53:40 ? 0:00 /usr/sfw/sbin/smbd –D Anyone any ideas?

    Read the article

  • Release Notes for 3/2/2012

    Here are the notes for today’s release: Added a progress indicator when saving issues. Added support for viewing CodePlex RSS feeds in Chrome. Deployed several bug fixes: Fixed an issue where the back button on Internet Explorer was not working as intended when browsing code. Fixed an issue where long commit comments would push the source control info box outside of the boundaries of the page. Fixed an issue where Internet Explorer users were not able to widen the frame of the source code browser until a file was selected. Fixed an issue where opening a source code file directly from a URL in Internet Explorer would cause the source code tree to be collapsed. Fixed an issue where adding a code snippet with long lines of text to a discussion thread using Internet Explorer would needlessly display a vertical scrollbar, limiting the amount of code visible. Fixed an issue where tabbing through some links would render them invisible. We deprecated support for embedding PreEmptive analytics statistics on the project statistics page. If you’re interested in collecting and reporting your own statistics, PreEmptive’s RunTime Intelligence Endpoint Starter Kit offers a good starting point for capturing data. Have ideas on how to improve CodePlex? Visit our ideas page! Vote for your favorite ideas or submit a new one. Got Twitter? Follow us and keep apprised of the latest releases and service status at @codeplex.

    Read the article

  • Internet Explorer 9 beta installation problem

    - by Eric
    I'm running Vista SP2, x64. Because I wanted to test out the IE9 beta, I downloaded the English-language installer for 64-bit Vista systems. Running the installer is fine until it starts downloading required updates. The progress bar doesn't get far before it completely stops moving. Then after 20 minutes to an hour, it will tell me that there's an update I have to install, but as soon as I click OK it sends me an error message, telling me that it can't go to the url of the update which is here. So I manually enter it into my browser, which prompts me to download a standalone update. After that's been downloaded and I run it, it tells me that the update does not apply to my system. I'd appreciate any help to solving my problem.

    Read the article

  • How to create a Service Connection Point for Exchange (Manually)

    - by Ionoxx
    I'm being cautious here. Before I remove anything I want to be able to put it back. I'm having issues with a domain joined computer that is using SCP to get exchange autodiscovery information. It's getting information for the now unused internal Exchange through SCP even through the profile is using Office 365 on another domain. According to this conversation, I can simply remove the object from Active Directory Sites and Services. I want to know how to add back in should this create more problems, or if we reinstate the Exchange server. Right clicking on the parent "autodiscover" node doesn't allow me to create a Service Connection Point. Will simply running the cmdlet "Set-ClientAccessServer -identity servername -AutodiscoverServiceInternalUri url" be enough to recreate the object? Thank you!

    Read the article

  • Setting correct Content-Type sent from Wordpress, on Apache server

    - by eoinoc
    I need help pointing me in the right direction for setting the ContentType returned by Apache for content produced by WordPress. I'm having trouble figuring out why WordPress is returning incorrect headers. Issue The specific problem is that our Wordpress blog pages are being downloaded as a file rather than displayed by Internet Explorer and Chrome v21. Content-Type: application/x-gzip is being returned by the server. I'm told that I should expect Content-Type: text/html. Background The URL is http://www.bitesizeirishgaelic.com/blog/.

    Read the article

  • Force firefox to open pages in a specific tab using command line

    - by user36306
    Hey Guys, Here's the challenge--I developed Softphone Screenpop PHP App that takes caller id info and searches for a match in our db, also allows us to collect call statistics. Great for our management but it's driving our reps nuts. We use firefox here and when our softphone pops to the external page, every time it opens in a new tab, the girls quickly get 5-10 open and it becomes confusing. Our softphone will also run command line. I wondering if there is a way to have a URL open in a certain tab. Otherwise does anyone have any other ideas? Thanks!

    Read the article

  • I think I disabled IE8's compatibility view

    - by malatio
    I have an install of IE8 that doesn't have the compatibility view button next to the url bar. I vaguely remember during install that there was a question about whether it was to be available or not, I may have clicked no. Or maybe that was a dream. There are no options that I can see about it being available or not. Anyone know how to enable it? I'm about to try re-install but I'm hoping I can fix it without re-installing. Version: 8.0.6001.18702 ; Update Versions: 0

    Read the article

  • Limiting to the ServerName in Apache2

    - by David
    I have 2 sites defined in my Apache2. Each one has a servername. For example: Server 1 (first in sites-enabled) responds to www.example.com Server 2 (second in sites-enabled) responds to www.example2.com Ok, the problem is when I type the server IP in the URL, the first server responds. How could I limit the response to only specifying its servername? I would like to block the IP calls. If that is not possible, I would like the second server to respond, not the first. I cannot change the order because there are aliases defined in the second server that would override the first server config.

    Read the article

  • Why is chrome slashing my HTTPS?

    - by John Isaacks
    In Chrome when I view my page the https in the URL has a red slash through it. When I click it, it says: Your connection to www.example.com is encrypted with 256-bit encryption. However, this page includes other resources which are not secure. ... I am not getting the mixed content warning in IE or FF. Is there a way to get Chrome to tell you exactly what/where is causing the issue? I am having trouble finding what is causing Chrome to throw a hissy fit. I searched my source for any http that does not contained https but am not seeing any. It would be nice if there is an easy way to pinpoint the issue.

    Read the article

  • Using Subdomains for Newly Regional Company

    - by Taylord22
    The company I work for is expanding their business to new territories. I've got a lot of stabilization to do in the region/state where we're one of the most well known companies of our kind. Currently, we have 3 distinct product lines which are currently distinguished by 3 separate URLS. This is affecting the user flow of our site, so we'd like to clean it up before launching our products into the various regions. The business has decided to grow into 5 new states (one state consisting of one county only) — none of which will feature all 3 products. Our homebase state is the only one that will have all 3 products this year. My initial thought was to use subdomains to separate out the regions, that way we could use a canonical tag to stabilize the root domain (which would feature home state content, and support content for all regions), and remove us from potential duplicate content penalization. Our product content will be nearly identical across the regions for the first year. I second guessed myself by thinking that it was perhaps better to use a "[product].root/region" URL instead. And I'm currently stuck by wondering if it was not better to build out subdomains for products and regions...using one modifier or the other as a funnel/branding page into the other. For instance, user lands on "region.root.com" and sees exactly what products we offer in that region. Basically, a tailored landing page. Meanwhile the bulk of the product content would actually live under "product.root.com/region/page". My head is spinning. And while searching for similar questions I also bumped into reference of another tag meant to be used in some similar cases to mine. I feel like there's a lot of risks involved in this subdomain strategy, but I also can't help but see the benefits in the user flow.

    Read the article

  • How do I generate and post XML in c# [migrated]

    - by user2922687
    I am a new to c# and faced with a similar problem. I need to generate and post XML to a URL but the parameter in the XML fields should dynamic getting inputs from a frontend app. This layout of the XML <soapenv:Envelope xmlns:soapenv="http://schemas.xmlsoap.org/soap/envelope/" xmlns:ret="http://xxx.xxx.com"> <soapenv:Header/> <soapenv:Body> <ret:Vend> <DestAccount >?</DestAccount> <Amount>?</Amount> <Msg>?</Msg> <SequenceNo>?</SequenceNo> <DealerNo>?</DealerNo> <Password>?</Password> </ret:Vend> </soapenv:Body> </soapenv:Envelope> Can anyone assist on how I can generate this in c#?

    Read the article

  • CentOS drive mapping? [on hold]

    - by DroidOS
    This is the first time I am posting on this particular StackExchange forum and I hope that I am using the right one for the present question. Briefly, this is what I need to do I am running a web service where users can, amongst other things, upload and store files on the server. What I want to do is to hive off user file storage to a different location so my server (CentOS 64 bit) can concentrate on what it does best - server side scripting and database management. As things stand all user files go into subdirectories in a folder called stash that lies above DOC_ROOT. What I would like to do is Transparently detect all attempts to read/write to stash/sub_folder and get/set file data on a remote server - ideally the latter would be one which replicates files like a CDN so they can be delivered from the closest/fastest location based on where the user's location. Even nicer would be if for all read accesses I could provide a URL that allows the user's browser to fetch the relevant file directly without having to funnel them via my server. I am a relative newbie when it comes to this sort of thing so I hope that I have phrased this question adequately well. From the little searching I done I gathered that WebDAV can be used to map drives to a different location on the web so perhaps that is a starting point. But if that will work I need to Establish how to get WebDAV up and running on my CentOS 64 bit server. Ideally, identify a service that allows this kind of file storage and provides an API I can use in my own scripting. I'd much appreciate any help with this.

    Read the article

< Previous Page | 512 513 514 515 516 517 518 519 520 521 522 523  | Next Page >