Map managed properties Path, OriginalPath to a custom property in SharePoint Online Search

We have a custom content type that we've added to a SharePoint list in each site collection to store site metadata. The content type has a URL field named Site URL that stores only the URL of the site collection. Now, when users search for content, the list entry appears and the URL displays the listing list URL? ID = on.

We want to map / prioritize the SiteURL custom URL property to the path, that is, the properties managed by OriginalPath so that the actual site collection URL maps to the search result. But it does not work that way. Properties fine assigns and below that is the value for both.

ows_q_URLH_MSCProjectSiteURL, ClientUrl, Basic: 11

Any ideas please suggest.


7 – Undefined property $ uid when running update.php

I had a small problem with running update.php. For example, after updating some modules, etc.

Whenever I start update.php, I get the following error:

Note: Undefined property: stdClass :: $ uid in user_uri () (line 194 of /home/master/public_html/modules/user/user.module).

This message does not point me to something more specific, so I'm not sure what to check here. I checked the Drupal logs for anything that appeared at the same time, and found nothing to do with it.
I ignored this message for a long time but I thought I would contact the community to see if anyone else saw it and if so, what they did.

gui – Change uipanel property values ​​from another .m file

I have created a GUI in Matlab that creates a MyGUI.fig file and a MyGUI.m file. If you use an existing Matlab script, we call it "MyCode.m". I would like to change the property values ​​in "MyGUI.fig".

The code basically compares a tag stored in an Excel file with a tag that is being transmitted serially, and then compares it. If they match, one field turns green, otherwise red.

This is the code for matlab r2017a:

delete(instrfind('Port', 'COM3')); 
tag = serial('COM3'); %check which port is used
BOX = char(zeros(2,14)); 
TrueValueData = 'C:MasterCodes.xlsx';
(~,~,TrueValMat) = xlsread(TrueValueData); % Creates matrix filled with the correct values, 
                                           % indexed by box, which is the first row
                                           % all proceeding rows are the master value

 function result(handles)
    for i=1:9223372036854775807
     if i>10 %first couple reads are filled with unicode nonsense, this skips that stage

     readData = fscanf(tag);
    if length(readData)>12

       BOX(str2num(readData(8)),1:14)= readData(11:24); % these numbers just give us what we want; 
                                                   % tags come in initially with some gobbledy-gook     
%      if(length(readData)>10) %if chip has been read
%      ReadChips

          if strcmp(TrueValMat{2,1}, BOX(1,:))
            set(handles.uipanel1, 'BackgroundColor', 'green');
             set(handles.uipanel1, 'BackgroundColor', 'red');

            if strcmp(TrueValMat{2,2}, BOX(2,:))
             set(handles.uipanel2, 'Backgroundcolor', 'green');
             set(handles.uipanel2, 'Backgroundcolor', 'red');
function uipanel1_Callback(hObject, eventdata, handles)
function uipanel2_Callback(hObject, eventdata, handles)

While running the script, MyGUI will be displayed and no errors will be output, but nothing happens when scanning the serial tags.

Any help is appreciated, I'm not very good at using Matlab.

The ViewModel property is NULL when sent to Main View – Asp.Net Core MVC

I have a ViewModel (FilialViewModel) that has a property of type PessoaViewModel. The main view is tied to FilialViewModel. For each submission, the property of the PersonViewModel type is NULL. This is because when I call PartialView Person, I pass FilialViewModel.PersonaViewModel … When sending, only the fields of FilialViewModel … are loaded.

public class FilialViewModel
    public int PessoaFilialId { get; set; }

    (DisplayName("Tipo de Filial"))
    (Required(ErrorMessage = "Escolha um Tipo de Filial"))
    public FilialTipo FilialTipo { get; set; }

    public PessoaViewModel PessoaViewModel { get; set; }


public class PessoaViewModel
    public int Id { get; set; }

    (Required(ErrorMessage = "Escolha uma Natureza"))
    public PessoaNatureza PessoaNatureza { get; set; }

Branch view:

@model Retaguarda.Application.ViewModels.Filial.FilialViewModel


@await Html.PartialAsync("~/Views/Pessoa/_Pessoa.cshtml", Model.PessoaViewModel)

Show person:

@model Retaguarda.Application.ViewModels.Pessoa.PessoaViewModel
@using Retaguarda.Domain.Enuns.Pessoa
    ViewData("Title") = "Pessoa";

@Html.HiddenFor(model => model.Id, new { @class = "hidden-id" })


(Authorize(Policy = "CanWriteFilialData"))
public IActionResult Create(FilialViewModel filialViewModel)
    if (!ModelState.IsValid) return View(filialViewModel);

    if (IsValidOperation())
        ViewBag.Sucesso = "Filial cadastrada!";
    // return Json(new { success = true, message = "Pessoa Excuída!" });
    return View(filialViewModel);


Enter image description here

Indexing – Does removing a Google Search Console property remove the Google index of the site?

Removing a property from Google Console removes only the site from Google Console.

I'm not sure what your goal is. However, you can use robots.txt to remove your site from Google, such as: B. by using …

User-agent: Googlebot
Disallow: /

… or all search engines

User-agent: *
Disallow: /

Each search engine has its own bot name. For example, Bing is a Bingbot.

User-agent: bingbot
Disallow: /

Robots.txt is a simple text file in the root of your website. It should be available as or

You can read about robots.txt at

For a list of major search engine bot / spider names, see the top search engine bot names.

Using the robots.txt file and the correct bot name is generally the quickest way to remove a website from a search engine. Once the search engine has read the robots.txt file, the site will be removed within about 2 days unless anything has changed lately. Google has deleted websites within 1-2 days. Each search engine is different and the responsiveness of each can vary. Please note that the major search engines react relatively quickly.

To address the comments.

Robots.txt is in fact used by search engines to know which pages need to be indexed. This is well known and well known and since 1994 a de facto standard.

That's how Google works

Google indexes, among other things, links, domains, URLs and page content.

The linkage table is used to discover new sites and pages, and to rank pages using the PageRank algorithm, which is based on the trusted network model.

The URL table is used as a link between links and pages.

If you know the SQL database schema,

The linkage table would look something like this:
Link text

The domain table would look something like this:
domain registrar

The URL table would look something like this:

The page table would look something like this:
page Title
Page Description
html page

The URL table is a linkage table between domains, links, and pages.

The page index is used to understand and index the contents of individual pages. Indexing is far more complicated than just an SQL table, but the illustration remains.

If Google follows a link, the link will be added to the link table. If the URL is not in the URL table, it is added to the URL table and passed to the fetch queue.

When Google retrieves the page, Google checks to see if the robots.txt file has been read, and if so, if it has been read within 24 hours. If the cached robots.txt data is older than 24 hours, Google will retrieve the robots.txt file again. If a page is restricted by robots.txt, Google will not index the page and will not remove the page from the index if it already exists.

When Google detects a restriction in robots.txt, it is sent to a queue for processing. The processing starts every night as batch processing. The pattern is matched against all URLs, and all pages are removed from the page table with the URL ID. The URL is kept for management.

Once the page has been retrieved, it will be inserted into the page table.

Any link in the join table that was not retrieved or restricted by robots.txt, or a bad link to a 4xx error, is called dangling links. And while PR can be calculated using the trust network theory for dangling links landing pages, PR can not be routed through these pages.

About 6 years ago, Google found it advisable to include drooping links in the SERPs. This happened when Google redesigned the index and systems to aggressively capture the entire Web. The idea was to present valid search results to users, even if the page was locked by the search engine.

URLs have very little or no semantic value.

Links have a certain semantic value, but this value remains low because semantic indexing prefers more text and does not perform well as a stand-alone element. Normally, the semantic value of a link is measured along with the semantic value of the source page (the page with the link) and the semantic value of the landing page.

As a result, a URL to a landing page of a dangling link may not have a good ranking at all. The exception are newly discovered links and pages. Typically, Google "tries" newly discovered links and pages within the SERPs by setting the PR values ​​high enough for them to be found and tested within the SERPs. Over time, PR and CTR are measured and adjusted to place links and pages where they should exist.

See ROBOTS.TXT DISALLOW: 20 years of mistakes to avoid, which also discusses the ranking I've described.

Listing links in the SERPs is wrong and many have complained about it. It pollutes the SERPs, for example, with broken links and links behind logins or paywalls. Google has not changed this approach. The ranking mechanisms, however, filter out the links from the SERPs and effectively remove them completely from the SERPs.

Remember that the indexing engine and the query engine are two different things.

Google recommends using noindex for pages that are not always possible or practical. However, I use noindex for very large sites that use automation. This can be impossible or at least cumbersome.

I had a website with millions of pages that I removed from the Google's index within a few days using the robots.txt file.

And while Google is against using the robots.txt file and using noindex, this is a much slower process. Why? Because Google uses a TTL-style metric in its index that determines how often Google visits this page. This can be a long period of time, which can last up to a year or more.

Using noindex does not remove the URL in the same way from SERPs as robots.txt. The final result remains the same. As it turns out, Noindex is actually no better than using the robots.txt file. Both produce the same effect, while the robots.txt file renders the results faster and in large quantities.

And this is partly the point of the robots.txt file. It is generally accepted that users block entire areas of their site using robots.txt or block bots from the site completely. This is more common than adding noindex to pages.

Removing an entire website using the robots.txt file is still the fastest way, even if Google does not like it. Google is neither God nor his website the New Testament. As much as Google tries, it still does not rule the world. Damn near, but not quite.

The claim that blocking a search engine using robots.txt prevents the search engine from seeing a noindex meta tag is utter nonsense and contradicts logic. You see this argument everywhere. Both mechanisms are in effect exactly the same, with the exception that one is much faster due to mass processing.

Keep in mind that the robots.txt standard was introduced in 1994, while the noindex meta tag was not yet adopted by Google in 1996. In the beginning, removing a page from a search engine meant using the file "robots.txt" file and stayed that way for a while. Noindex is just an extension of the existing process.

Robots.txt remains the # 1 mechanism to limit what a search engine will index and likely to do while I'm alive. (Be careful when crossing the road, no more skydiving for me!)

javascript – JS ERROR The property & # 39; childNodes & # 39; by undefined visual studio 2017 ASP.NET can not be read

I am developing a project in Visual Studio 2017 in ASP.NET and have a problem with a function in js in a view.
I know why this problem usually occurs, but I can not see it here.
The function is the following:

var columnas = document.getElementsByClassName("nombresgrups");
            for (i = 0; i < columnas.length; i++) {

                for (let a = 0; grupos.length; a++) {
                    if (columnas(i).getAttribute("id") == grupos(a).childNodes(0).nodeValue)
                        console.log("exito: ");

The error starts

if (columnas(i).getAttribute("id") == grupos(a).childNodes(0).nodeValue)

The story is that the protocol before the error does it correctly and is the same value I'm trying to access (groups (a) .childNodes (0) .nodeValue).
This log is written once and then outputs the error. It enters the loop, gives me the value, and then outputs the error.

The variable groups come from reading an XML and at this point it has an element.
I have attached a screenshot of the console.

I have attached a screenshot of the console.

dnd 5e – Can the Spell of Motion spell prevent the Gibbering Mouther's Aberrant Ground property from lowering a creature's speed to 0 in the event of a failed Rescue?

The freedom of movement The spell only prevents difficult terrain from affecting our target Move, The saving throw happens anyway

The freedom of movement Spell states:

During the entire time, the movement of the target remains unaffected by difficult terrain. Spells and other magical effects can not reduce the target's speed or cause the target to be paralyzed or held back.

From this we can conclude what the spell does:

  1. Difficult terrain does not affect your Move. Note that this does not say anything about your speed, hit points or other effects on difficult terrain.

  2. Spells and magical effects can not reduce your speed.

  3. Spells and magical effects can not paralyze or hold you back.

Gibbering Mouther's Aberrant Ground feature is not magical, so the last two points do not apply. The function causes the following:

The ground within a radius of 10 feet around the mountain is doughy terrain. Each creature starting their turn in this area must receive a DC Save Throw of 10 or reduce their pace to 0 by the beginning of their next turn.

  1. Ground within a 10 foot radius becomes difficult terrain.

  2. If a creature starts its turn within a 10 foot radius, it must make a save throw. If you do not perform this save, your speed will become 0.

The function never says that we automatically (or should not) make the saving throw if we are immune to the normal effect of difficult terrain. Since freedom of movement only helps to prevent changes to ours Move However, this is a change in our speed, which is still normal for us.

Likewise with a spell like tip growthcreating a damaging area of ​​difficult terrain,freedom of movement will not stop us somehow from being damaged; it only prevents the area from costing extra Move,

Can I add custom metrics and dimensions to the Google Analytics Web Plus App property?

I have a web property that allows me to add custom metrics and dimensions to track them in my web app: Admin -> Property Settings -> custom definitions

However, I've created a new Web + App property in Google Analytics and I can not seem to find a way to add custom metrics and dimensions.

Is there a way to add custom metrics and dimensions to Web + App properties?

Google Analytics – Can I add custom metrics and dimensions to the WebPlus app property?

I have a web property that allows me to add custom metrics and dimensions to track them in my web app: Admin -> Property Settings -> custom definitions

However, I've created a new Web + App property in Google Analytics and I can not seem to find a way to add custom metrics and dimensions.

My question is: is there a way to add custom metrics and dimensions to Web + App properties?