Search Results

Search found 9727 results on 390 pages for 'llblgen pro'.

Page 240/390 | < Previous Page | 236 237 238 239 240 241 242 243 244 245 246 247  | Next Page >

  • Customer escalated to a claim without sending the item back? [closed]

    - by kavoir.com
    She claims that she has sent the item back but for over 1 month I haven't received it. I don't know where I can find the tracking number so I don't know if she really sent it or not. Now she has escalated the dispute to claim and PayPal is asking me for information. The reason is "Not as described". So how do I respond to this? I mean, in the dispute, we agreed that I'll issue a full refund as soon as I receive the package she return to us but we never received this package that she claimed to have sent back. Now she's escalating this to a claim and PayPal is asking me for documents. How can I provide any documents that would prove she hasn't sent the package? Thanks!

    Read the article

  • Making HTML5 videos stored on AWS S3 **difficult** to download (because I cant make it impossible)

    - by Jimmery
    I am building a website that hosts video's stored on AWS's S3 service. The videos are played thru a HTML5 player we have built. Ive just been asked to make sure "nobody can steal our video's". Now I know that if you really don't want something stolen, don't put it up on the internet. However I just need to secure these videos as good as possible, the videos need to at the very least resist someone going thru the source code and trying to download them manually. One option available to me is to completely rebuild the video player in flash. This is not ideal, for several reasons, notably because I would also then have to build an App for mobile devices to be able to view this site. So I am looking for other options. I have heard about using a token to make the file available only during certain times. I have heard of using a separate file to serve the videos that sits between the HTML5 page and the video file. I am also having a look at IAM, the Secure AWS Access Control, in the hopes AWS can solve this problem for me. Can anyone here recommend any of these options? Or perhaps suggest other options available to me? Any help would be greatly appreciated.

    Read the article

  • How to set up a local webDAV server for use with Goodreader on the ipad? [migrated]

    - by confused-about-webdav
    I need to know how to set up a local webDav server on my PC so that goodreader on ipad can automatically sync with it over the local wifi network? I am really a rookie when it comes to setting up a web server and have tried various guides on the internet. I tried setting up a webdav server using IIS and forwarded the required ports and enabled webdav publishing but goodreader can find it in the local wifi network automatically nor is it able to connect even after manually entering the credendtials. So i'll be really greateful if someone who has successfully setup a webdav server for use with goodreader can point me on how to do it?

    Read the article

  • Problem in domain name of site in google search [on hold]

    - by Jayadratha Mondal
    My domain provider dont supports advanced DNS at this moment. So I have used iframe to forward to my webserver. Suppose my domain abc.com have a simple html which is opening xyz.com via Iframe. Google was showing abc.com on the search. But from last two days it is showing xyz.com, other things like description, name of the site are ok only the domain changed to xyz.com, but I want to show abc.com. When I'm typing the full domain then google showing the domain as I want. But If I type a part or any keyword then it showing that xyz.com. Anyone have any idea how to show abc.com again? EDIT: I dont know the domain name provider because its not mine. When I'm trying to set new A record its showing that "you need advanced dns to do this currently you are using simple DNS". I'm doing this from cpanel. Normally as per I know there are 3 sections. 1 domain name 2 A record 3Cname. But I dont have any domain name section.

    Read the article

  • Css clicking problem [migrated]

    - by Shyghar
    I'm building a strange div shaped structure and I need a hint to resolve a clicking problem. This is a jsfiddle to show you the issue. The structure for each element is: <div class="views-row"> <div class="diamonds-container"> CONTENT </div> </div> I have a onclick() event on .diamonds-container but the .views-row div of the next element [with red or blue background..] go over the container and stop the click event on it. I tryed to play with the z-index but I didn't have the expected result. How can I achieve this structure with a correct click event on diamonds-containers ? I think I can track the .views-row click with javascript and trigger manually a click on the previous diamonds-container but this will be my final option. How can I achieve this without javascript?

    Read the article

  • Cloud proxying service

    - by ChristopherJ
    I have an app that mashes up images from Bing image search, it's hosted on Heroku written in rails. The app is client side in javascript, so the mashup is done on an html5 canvas - this means though that if I fetch the images direct from the Bing server, the canvas gets dirty and I can't save it. As a quick work around, i have set up a route on my rails app that simply proxies the request to Bing and passes the result back through. Obviously this is a very poor performance solution and will eat up my dynos very quickly. Can anyone suggest a more suitable option? At the moment I'm thinking maybe Amazon EC2 with apache mod_rewrite rules would be better performing and more cost effective. Is there a cloud service (or an app I could deploy to a cloud service) that would be more appropriate for proxying requests for me so that my javascript can fetch the images without dirtying the canvas?

    Read the article

  • google analytics statistics

    - by colmcq
    I am compiling a report for a client using google analytics. I have observed that the client has unusually good page view times (5 mins) and excellent bounce rates (<25%). I need to reference research data that validates my assertion that these figures are excellent compared to an industry standard (the industry is ecommerce and gaming). Can you direct me to any published research data that specifies normal bounce rates and page view times for this industry?

    Read the article

  • Two different domain for specific languages pointing to one site

    - by user25599
    I developing a client's blog and he needs it to be bilingual (english and spanish). Now what he wants is that users can get to the content based on the domain e.g. Jhon enters www.domain.com and he gets the english version and Juan enters www.elsenordominio.com to get the spanish version. All content will be validated by php so the users and search engines only reads the domain related language. What do i need to use header re-direct or 301? is it bad for seo? will google will google penalize me? I hope you guys can help me and forgive me if my english is not good.

    Read the article

  • Forwarding naked domain to www [duplicate]

    - by mravec
    This question already has an answer here: Why is www. working but no-www is not? 1 answer I successfully mapped my www.mydomain.com to my Google site (as described here). However I would like to forward also my naked domain (mydomain.com) to end up in www.mydomain.com. I assume I should modify 'A' DNS record in my provider but what IP should I point it to?

    Read the article

  • Subdomains vs. URL Path in shareable links

    - by Adam Matan
    I am building a web application for questions and answers. Each question/answer page has all the required metadata for Facebook and Twitter, and we encourage users to share these pages. I have a dilemma regarding the shared link structure: Option 1 - subdomains Use a questions.example.com and answers.example.com, followed by an ID and optional text. The text is ignored by the request, which only takes the id into account. http://questions.example.com/<question_id>/<question_text> http://questions.example.com/12345/how-long-is-the-queue # Example http://q.example.com/12345 # Example Option 2 - URL path This is the format used by stackoverflow.com and trello.com: http://example.com/questions/<question_id>/<question_text> http://example.com/questions12345/how-long-is-the-queue # Example http://example.com/q/12345 # Example Server-wise, I can easily do both - I have a wildcard SSL certificate and Apache/NGinx configuration is pretty straightforward. Which option - subdomains or URL path - is preferred for shareble links?

    Read the article

  • Why am I getting this message "Some important page is blocked by robots.txt"?

    - by Rounak
    My site's URL is: www.hackinguniverse.org From some day in Google Webmaster tool a message is showing that says "Some important page is blocked by robots.txt". My robots.txt is: User-agent: Mediapartners-Google Disallow: User-agent: * Disallow: /search Allow: / Sitemap: http://www.hackinguniverse.org/feeds/posts/default?orderby=updated Now for other information - I hosted this website on blogger. I have some other sitemaps too where I had only included "/Search" in Disallow like in this robots.txt file. But those sites are ok. I mean no message is showing on those. So why am I getting that message telling that I have blocked some important page via robot.txt?

    Read the article

  • How to make Google show my site in search result like the following image? [closed]

    - by Samik Chattopadhyay
    Possible Duplicate: What are the most important things I need to do to encourage Google Sitelinks? Currently Google is displaying my site (http://layzend.info) like this in search result, only the link and meta description without any internal page links - But I want to be the search result like the following where the internal links are also displayed - How is it possible? Please help me to make my site more SEO friendly.

    Read the article

  • .htaccess two different rules but only one per time

    - by dragon112
    I'm rather new to the whole .htaccess thing and I'm using the following right now to use 'pretty url's': <IfModule mod_rewrite.c> RewriteEngine On RewriteCond %{REQUEST_FILENAME} !-d RewriteCond %{REQUEST_FILENAME} !-f RewriteRule ^(.*)$ index.php?path=$1 [NS,L] </IfModule> Now i found my website a bit slow and decided to start gzipping my CSS files thru a php script I found somewhere on the web (the website). For this to work I need to rewrite the url to open the correct php file. That would look something like this: RewriteRule ^(.*).css$ /csszip.php?file=$1.css [L] But I only want the first to happen when the second doesn't and vice versa. In other words i'd like something like this: <IfModule mod_rewrite.c> RewriteEngine On if request doesn't contain .css do RewriteCond %{REQUEST_FILENAME} !-d RewriteCond %{REQUEST_FILENAME} !-f RewriteRule ^(.*)$ index.php?path=$1 [NS,L] else do RewriteRule ^(.*).css$ /csszip.php?file=$1.css [L] </IfModule> Can anyone help me with the proper code or a place where i can find a way to use some kind of conditional statement in htaccess files? Thanks in Advance!:)

    Read the article

  • How to prevent a search engines from indexing a section of a page?

    - by BrunoLM
    I have many pages with lots of text in it. But I will always have two sections of text and I want to prevent one section from appearing in search results, the other section must be indexed. <p class="please-index-me">text</p> <p class="get-out">never index me please</p> I thought that maybe if I load the "please don't index me text" with Javascript maybe search engines wouldn't look for it. But I am not sure it would work and this is not really nice. I was wondering if there is a way to tell search engines "hey, this text you can't grab, move on". So, is there a way to do it?

    Read the article

  • Impact of migrating home page from HTTP to HTTPS on search results

    - by 2Stroke SEO
    I've had to change one of my site's home page to HTTPS from HTTP. I had plenty of links coming into the HTTP page and was performing well in Google against many of my targeted search phrases. I did a 301 redirect from the HTTP page to transfer the link juice to the HTTPS page (and to prevent duplicate content issues) but my search rankings have tanked which indicates no link juice has been transferred. My PageRank has vanished - which I'd expect - but I'm really surprised that the SERP rankings fell off the face of the earth. Anyone have any ideas how I can recover from this? I've waited a couple of months since the changes took effect just in case Google was taking time to check it out.

    Read the article

  • XCart Skip Checkout Steps

    - by user900553
    I am using xcart. In the checkout process, skip the step that appears after a customer clicks the add to cart button on the product page. It takes them to a page that shows the product and the text Added to your shopping cart. I want to take them directly to the view cart page, right before the checkout page. I also tried: Enable Redirect customer to cart after adding a product option in X-Cart admin back-end -Settings - General settings. Still not working. Anybody have Idea about this?

    Read the article

  • Do I really need to remove special characters in a URL?

    - by anarchoi
    I have an FTP account shared with friends where we upload underground music albums and then we use the links to share the downloads in a music forum. The problem is that the album names are in french so there is a lot of special characters in the name. So the URL looks like http://www.mydomain.com/downloads/Some Band - En français avec des caractères spéciaux (2013) [7'' EP].zip For me it works perfectly and I can download the file by using this URL, but I have read everywhere that special chars are bad in URL. Is there any reason why I must remove the special characters or encode the URL? Is everyone able to access a URL with special characters or will some older browsers not be able to download the files? I really don't care about SEO or anything else. I just want the download links to work for everyone. Since the files are uploaded through FTP, I can't use PHP to remove the special characters with a regex, so I really don't know what to do.

    Read the article

  • Redirect subdomain to local pc

    - by user1188570
    I have a home webserver which is constatly running. Is it possible to create a subdomain which would redirect traffic to another local pc? For example I have 1 Server and one notebook(with webserver installed for developing) Now I can access to notebook only from local network with IP. Server is also hosting domain example.com. Now I would like to visit laptop.example.com, which would be my laptop.

    Read the article

  • Image captions and wrapping [migrated]

    - by Charles
    What's the best way to add a caption below an image? The image and its caption will be floated right, and the text on the caption needs to wrap -- a 200x200px image shouldn't have a caption of width 800px. I would strongly prefer a solution that allows me to update images (with different widths) without changing the CSS or markup. For reasons beyond my control the image itself will also be floated right, but this should not be too problematic.

    Read the article

  • How would a search engine see url encoded characters?

    - by K20GH
    I've got my URL however some of the strings would contain &. Obviously I can't use them as best practice so I've replaced them with +. However if I encoded my & instead it would become %26. How would a search engine see that? Would it see %26 as a & so still bring back the URL or would it just see it as a %26? ie. Would www.example.com/sweet?m&m show as that, or would they see it as www.example.com/sweet?m%26m

    Read the article

  • Blank lines between sourcecode [closed]

    - by manix
    I'm so confused with a strange behaviour. Actually I have edited some php files remotely with my PhpDesigner8 (a php editor). Everything goes right, but when my teammates reopen the files that I have edited the source code have blank lines like below: class AdminController extends Controller { function __construct() { parent::__construct(); if (!$this->session->can_admin()) { show_error('Solo para administradores.'); } $this->load->library('backend'); } } Instead of class AdminController extends Controller { function __construct() { parent::__construct(); if (!$this->session->can_admin()) { show_error('Solo para administradores.'); } $this->load->library('backend'); } } Did you have experience these kinds of problems?

    Read the article

  • SEO consideration for duplicate sites

    - by Malk
    I am building a brochure-ware website for a company that sells products all across the world. They need the site to ask the user what region they are in before using the site; there are 5 regions. This is because there are different products offered to different regions and each region may or may not want to customize their own content. However, at launch and likely forever, most of the pages will be the exact same minus what is listed in the footer and in the product selection menu. My question is how should I structure the sitemap for this site for best SEO? Should I be concerned with duplicate content penalties and/or cannibalizing the site's presence on the SERP? Some considerations: The client wants to be able to print links directly to regional specific content bypassing any prompt for the user to select a region (to ensure they land on the target page). The client cannot have a 'default' region so the user must have a region specified "Clean" urls are important, but there is wiggle room The client does not want each region to have its own domain There will be a link on the page to allow users to specify a different region The client is not concerned with localization ...at this time Some products are available in multiple regions A quick list of options I am considering: www.site.com/region/page region.site.com/page www.site.com/page?region (no cookie, pages require the parameter. If visited without; the user must select a region) www.site.com/page (using cookie and a splash screen if needed; could pass parameter in to set the region for direct linking) Thanks in advance for your advice.

    Read the article

  • What is the best and cheapest CMS for an E-Commerce site?

    - by kf0l
    My client needs a site customers can use to pay the membership fees and join. It should also track when old members need to pay again. There are some members that will not want to pay online and they want to be able to use the site to manage all payments. The CMS must be cheap to host online, potentially with something like GitHub/Heroku. I am experienced with PHP and RoR. I am using a CMS so once I am done with this project it is easy for them to update. After reading through different posts I think WordPress with an e-commerce Plugin may be the right solution. Spree and WebGUI also looks promising. Thanks for the help.

    Read the article

  • Using Ajax #! for Google but site is not being crawled any more

    - by user28231
    We used to have all links in a normal way (thus with parameters?), we changed to Ajax loading cause we installed a music player. Here's an example: Old links: http://stereofox.com/post.php?idPost=5326 New links: http://stereofox.com/post.php#!idPost=5326 The snapshot you can get it by adding _escaped_fragment_= after the ?. All of these have the same content, and this image says what has happend to the site since we changed the linking system.

    Read the article

< Previous Page | 236 237 238 239 240 241 242 243 244 245 246 247  | Next Page >